var/home/core/zuul-output/0000755000175000017500000000000015113251646014531 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015113263306015471 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005230531215113263300017670 0ustar rootrootDec 01 08:40:23 crc systemd[1]: Starting Kubernetes Kubelet... Dec 01 08:40:23 crc restorecon[4744]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:23 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 08:40:24 crc restorecon[4744]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 08:40:24 crc restorecon[4744]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 01 08:40:24 crc kubenswrapper[4873]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 01 08:40:24 crc kubenswrapper[4873]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 01 08:40:24 crc kubenswrapper[4873]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 01 08:40:24 crc kubenswrapper[4873]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 01 08:40:24 crc kubenswrapper[4873]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 01 08:40:24 crc kubenswrapper[4873]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.215713 4873 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223051 4873 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223111 4873 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223123 4873 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223132 4873 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223143 4873 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223152 4873 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223161 4873 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223170 4873 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223179 4873 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223187 4873 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223195 4873 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223205 4873 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223214 4873 feature_gate.go:330] unrecognized feature gate: Example Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223223 4873 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223231 4873 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223239 4873 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223247 4873 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223254 4873 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223262 4873 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223270 4873 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223278 4873 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223285 4873 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223293 4873 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223301 4873 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223309 4873 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223317 4873 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223324 4873 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223332 4873 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223339 4873 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223347 4873 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223358 4873 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223368 4873 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223377 4873 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223399 4873 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223408 4873 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223416 4873 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223424 4873 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223433 4873 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223443 4873 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223452 4873 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223461 4873 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223469 4873 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223477 4873 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223485 4873 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223492 4873 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223500 4873 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223508 4873 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223515 4873 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223523 4873 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223531 4873 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223539 4873 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223547 4873 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223554 4873 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223561 4873 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223569 4873 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223577 4873 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223585 4873 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223593 4873 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223601 4873 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223609 4873 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223617 4873 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223625 4873 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223633 4873 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223640 4873 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223648 4873 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223661 4873 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223671 4873 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223682 4873 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223691 4873 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223699 4873 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.223710 4873 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.223875 4873 flags.go:64] FLAG: --address="0.0.0.0" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.223894 4873 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.223917 4873 flags.go:64] FLAG: --anonymous-auth="true" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.223937 4873 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.223948 4873 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.223958 4873 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.223971 4873 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.223982 4873 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.223991 4873 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224001 4873 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224011 4873 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224058 4873 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224067 4873 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224077 4873 flags.go:64] FLAG: --cgroup-root="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224085 4873 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224095 4873 flags.go:64] FLAG: --client-ca-file="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224103 4873 flags.go:64] FLAG: --cloud-config="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224113 4873 flags.go:64] FLAG: --cloud-provider="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224121 4873 flags.go:64] FLAG: --cluster-dns="[]" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224146 4873 flags.go:64] FLAG: --cluster-domain="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224155 4873 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224164 4873 flags.go:64] FLAG: --config-dir="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224173 4873 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224183 4873 flags.go:64] FLAG: --container-log-max-files="5" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224194 4873 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224203 4873 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224228 4873 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224238 4873 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224248 4873 flags.go:64] FLAG: --contention-profiling="false" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224257 4873 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224265 4873 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224277 4873 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224286 4873 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224306 4873 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224315 4873 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224324 4873 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224333 4873 flags.go:64] FLAG: --enable-load-reader="false" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224342 4873 flags.go:64] FLAG: --enable-server="true" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224351 4873 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224373 4873 flags.go:64] FLAG: --event-burst="100" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224382 4873 flags.go:64] FLAG: --event-qps="50" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224392 4873 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224401 4873 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224410 4873 flags.go:64] FLAG: --eviction-hard="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224421 4873 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224430 4873 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224439 4873 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224460 4873 flags.go:64] FLAG: --eviction-soft="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224470 4873 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224479 4873 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224487 4873 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224496 4873 flags.go:64] FLAG: --experimental-mounter-path="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224505 4873 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224514 4873 flags.go:64] FLAG: --fail-swap-on="true" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224523 4873 flags.go:64] FLAG: --feature-gates="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224534 4873 flags.go:64] FLAG: --file-check-frequency="20s" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224543 4873 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224552 4873 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224565 4873 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224574 4873 flags.go:64] FLAG: --healthz-port="10248" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224583 4873 flags.go:64] FLAG: --help="false" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224623 4873 flags.go:64] FLAG: --hostname-override="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224632 4873 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224643 4873 flags.go:64] FLAG: --http-check-frequency="20s" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224654 4873 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224663 4873 flags.go:64] FLAG: --image-credential-provider-config="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224672 4873 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224681 4873 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224690 4873 flags.go:64] FLAG: --image-service-endpoint="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224700 4873 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224709 4873 flags.go:64] FLAG: --kube-api-burst="100" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224718 4873 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224728 4873 flags.go:64] FLAG: --kube-api-qps="50" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224737 4873 flags.go:64] FLAG: --kube-reserved="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224746 4873 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224755 4873 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224765 4873 flags.go:64] FLAG: --kubelet-cgroups="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224773 4873 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224782 4873 flags.go:64] FLAG: --lock-file="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224791 4873 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224801 4873 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224810 4873 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224824 4873 flags.go:64] FLAG: --log-json-split-stream="false" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224846 4873 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224856 4873 flags.go:64] FLAG: --log-text-split-stream="false" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224865 4873 flags.go:64] FLAG: --logging-format="text" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224873 4873 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224883 4873 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224892 4873 flags.go:64] FLAG: --manifest-url="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224901 4873 flags.go:64] FLAG: --manifest-url-header="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224915 4873 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224925 4873 flags.go:64] FLAG: --max-open-files="1000000" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224936 4873 flags.go:64] FLAG: --max-pods="110" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224945 4873 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224954 4873 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224964 4873 flags.go:64] FLAG: --memory-manager-policy="None" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224973 4873 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224982 4873 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.224991 4873 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225001 4873 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225044 4873 flags.go:64] FLAG: --node-status-max-images="50" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225053 4873 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225063 4873 flags.go:64] FLAG: --oom-score-adj="-999" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225072 4873 flags.go:64] FLAG: --pod-cidr="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225081 4873 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225095 4873 flags.go:64] FLAG: --pod-manifest-path="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225103 4873 flags.go:64] FLAG: --pod-max-pids="-1" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225113 4873 flags.go:64] FLAG: --pods-per-core="0" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225122 4873 flags.go:64] FLAG: --port="10250" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225131 4873 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225140 4873 flags.go:64] FLAG: --provider-id="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225149 4873 flags.go:64] FLAG: --qos-reserved="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225157 4873 flags.go:64] FLAG: --read-only-port="10255" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225168 4873 flags.go:64] FLAG: --register-node="true" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225177 4873 flags.go:64] FLAG: --register-schedulable="true" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225187 4873 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225202 4873 flags.go:64] FLAG: --registry-burst="10" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225211 4873 flags.go:64] FLAG: --registry-qps="5" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225220 4873 flags.go:64] FLAG: --reserved-cpus="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225241 4873 flags.go:64] FLAG: --reserved-memory="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225253 4873 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225262 4873 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225271 4873 flags.go:64] FLAG: --rotate-certificates="false" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225279 4873 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225288 4873 flags.go:64] FLAG: --runonce="false" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225297 4873 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225306 4873 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225315 4873 flags.go:64] FLAG: --seccomp-default="false" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225324 4873 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225333 4873 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225342 4873 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225352 4873 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225361 4873 flags.go:64] FLAG: --storage-driver-password="root" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225370 4873 flags.go:64] FLAG: --storage-driver-secure="false" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225380 4873 flags.go:64] FLAG: --storage-driver-table="stats" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225388 4873 flags.go:64] FLAG: --storage-driver-user="root" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225398 4873 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225408 4873 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225417 4873 flags.go:64] FLAG: --system-cgroups="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225425 4873 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225439 4873 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225448 4873 flags.go:64] FLAG: --tls-cert-file="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225457 4873 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225475 4873 flags.go:64] FLAG: --tls-min-version="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225483 4873 flags.go:64] FLAG: --tls-private-key-file="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225500 4873 flags.go:64] FLAG: --topology-manager-policy="none" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225509 4873 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225519 4873 flags.go:64] FLAG: --topology-manager-scope="container" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225530 4873 flags.go:64] FLAG: --v="2" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225543 4873 flags.go:64] FLAG: --version="false" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225555 4873 flags.go:64] FLAG: --vmodule="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225566 4873 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.225576 4873 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.225790 4873 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.225800 4873 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.225820 4873 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.225831 4873 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.225840 4873 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.225849 4873 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.225857 4873 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.225865 4873 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.225873 4873 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.225881 4873 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.225889 4873 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.225896 4873 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.225904 4873 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.225912 4873 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.225920 4873 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.225928 4873 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.225935 4873 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.225943 4873 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.225951 4873 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.225959 4873 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.225968 4873 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.225976 4873 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.225984 4873 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.225991 4873 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.226003 4873 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.226012 4873 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.226044 4873 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.226052 4873 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.226060 4873 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.226068 4873 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.226076 4873 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.226085 4873 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.226093 4873 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.226101 4873 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.226109 4873 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.226117 4873 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.226125 4873 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.226133 4873 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.226154 4873 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.226162 4873 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.226171 4873 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.226180 4873 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.226190 4873 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.226199 4873 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.226208 4873 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.226216 4873 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.226224 4873 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.226235 4873 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.226246 4873 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.226255 4873 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.226267 4873 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.226278 4873 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.226288 4873 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.226297 4873 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.226306 4873 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.226315 4873 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.226327 4873 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.226338 4873 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.226346 4873 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.226354 4873 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.226362 4873 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.226373 4873 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.226383 4873 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.226392 4873 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.226400 4873 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.226408 4873 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.226417 4873 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.226425 4873 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.226433 4873 feature_gate.go:330] unrecognized feature gate: Example Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.226441 4873 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.226449 4873 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.226464 4873 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.239363 4873 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.239425 4873 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.239585 4873 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.239607 4873 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.239619 4873 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.239814 4873 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.239824 4873 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.239834 4873 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.239842 4873 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.239851 4873 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.239860 4873 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.239869 4873 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.239878 4873 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.239887 4873 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.239895 4873 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.239904 4873 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.239912 4873 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.239920 4873 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.239928 4873 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.239936 4873 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.239944 4873 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.239952 4873 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.239959 4873 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.239967 4873 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.239975 4873 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.239983 4873 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.239991 4873 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.239999 4873 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240007 4873 feature_gate.go:330] unrecognized feature gate: Example Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240037 4873 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240045 4873 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240053 4873 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240061 4873 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240069 4873 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240076 4873 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240084 4873 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240094 4873 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240102 4873 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240110 4873 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240118 4873 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240125 4873 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240133 4873 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240144 4873 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240155 4873 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240164 4873 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240172 4873 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240183 4873 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240193 4873 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240204 4873 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240213 4873 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240222 4873 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240231 4873 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240239 4873 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240247 4873 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240255 4873 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240263 4873 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240273 4873 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240283 4873 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240292 4873 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240301 4873 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240309 4873 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240317 4873 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240325 4873 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240333 4873 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240340 4873 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240348 4873 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240356 4873 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240364 4873 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240372 4873 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240380 4873 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240387 4873 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240395 4873 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240404 4873 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.240418 4873 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240760 4873 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240773 4873 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240781 4873 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240790 4873 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240800 4873 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240807 4873 feature_gate.go:330] unrecognized feature gate: Example Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240815 4873 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240823 4873 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240831 4873 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240840 4873 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240848 4873 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240855 4873 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240863 4873 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240871 4873 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240879 4873 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240886 4873 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240894 4873 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240902 4873 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240909 4873 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240920 4873 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240930 4873 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240938 4873 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240946 4873 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240955 4873 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240964 4873 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240971 4873 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240980 4873 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240987 4873 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.240995 4873 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.241003 4873 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.241011 4873 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.241041 4873 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.241049 4873 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.241059 4873 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.241069 4873 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.241078 4873 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.241086 4873 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.241094 4873 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.241101 4873 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.241109 4873 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.241117 4873 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.241125 4873 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.241133 4873 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.241140 4873 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.241149 4873 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.241157 4873 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.241165 4873 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.241173 4873 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.241180 4873 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.241188 4873 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.241196 4873 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.241204 4873 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.241212 4873 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.241219 4873 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.241227 4873 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.241235 4873 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.241245 4873 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.241255 4873 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.241264 4873 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.241272 4873 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.241280 4873 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.241289 4873 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.241297 4873 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.241304 4873 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.241312 4873 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.241320 4873 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.241330 4873 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.241340 4873 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.241350 4873 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.241358 4873 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.241367 4873 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.241380 4873 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.241874 4873 server.go:940] "Client rotation is on, will bootstrap in background" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.247178 4873 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.247365 4873 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.248451 4873 server.go:997] "Starting client certificate rotation" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.248524 4873 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.248890 4873 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-21 23:23:10.064152904 +0000 UTC Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.249118 4873 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.257329 4873 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 01 08:40:24 crc kubenswrapper[4873]: E1201 08:40:24.260225 4873 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.195:6443: connect: connection refused" logger="UnhandledError" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.260762 4873 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.272533 4873 log.go:25] "Validated CRI v1 runtime API" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.304744 4873 log.go:25] "Validated CRI v1 image API" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.307554 4873 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.311057 4873 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-01-08-35-54-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.311103 4873 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.339010 4873 manager.go:217] Machine: {Timestamp:2025-12-01 08:40:24.33677531 +0000 UTC m=+0.238883899 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654132736 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:27c66b0b-2450-400c-b6ea-cdb9cbbf95ad BootID:963d6124-09f6-48b8-b38c-854877aa92e3 Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730829824 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827068416 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:35:e3:ff Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:35:e3:ff Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:0c:e6:d8 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:5e:71:98 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:c3:18:75 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:38:f6:6b Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:4b:b9:d1 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:36:af:37:31:4b:d9 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:ba:6a:45:12:aa:20 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654132736 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.339551 4873 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.339780 4873 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.340852 4873 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.341207 4873 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.341274 4873 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.341607 4873 topology_manager.go:138] "Creating topology manager with none policy" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.341627 4873 container_manager_linux.go:303] "Creating device plugin manager" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.342143 4873 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.342209 4873 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.342635 4873 state_mem.go:36] "Initialized new in-memory state store" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.342795 4873 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.343813 4873 kubelet.go:418] "Attempting to sync node with API server" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.343846 4873 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.343885 4873 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.343907 4873 kubelet.go:324] "Adding apiserver pod source" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.343925 4873 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.346553 4873 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.347399 4873 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.347400 4873 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.195:6443: connect: connection refused Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.347415 4873 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.195:6443: connect: connection refused Dec 01 08:40:24 crc kubenswrapper[4873]: E1201 08:40:24.347493 4873 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.195:6443: connect: connection refused" logger="UnhandledError" Dec 01 08:40:24 crc kubenswrapper[4873]: E1201 08:40:24.347502 4873 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.195:6443: connect: connection refused" logger="UnhandledError" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.349108 4873 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.350079 4873 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.350125 4873 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.350142 4873 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.350185 4873 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.350209 4873 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.350223 4873 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.350237 4873 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.350261 4873 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.350279 4873 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.350295 4873 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.350316 4873 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.350331 4873 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.350397 4873 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.351561 4873 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.195:6443: connect: connection refused Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.352275 4873 server.go:1280] "Started kubelet" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.352533 4873 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.352706 4873 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.353900 4873 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 01 08:40:24 crc systemd[1]: Started Kubernetes Kubelet. Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.356983 4873 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.357325 4873 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.357815 4873 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 06:55:02.874915873 +0000 UTC Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.358162 4873 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 358h14m38.516764259s for next certificate rotation Dec 01 08:40:24 crc kubenswrapper[4873]: E1201 08:40:24.357684 4873 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.195:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187d0aba51d53db0 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-01 08:40:24.352226736 +0000 UTC m=+0.254335305,LastTimestamp:2025-12-01 08:40:24.352226736 +0000 UTC m=+0.254335305,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.358808 4873 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.358858 4873 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.358944 4873 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.361729 4873 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.195:6443: connect: connection refused Dec 01 08:40:24 crc kubenswrapper[4873]: E1201 08:40:24.361948 4873 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.195:6443: connect: connection refused" logger="UnhandledError" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.362294 4873 factory.go:55] Registering systemd factory Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.362462 4873 factory.go:221] Registration of the systemd container factory successfully Dec 01 08:40:24 crc kubenswrapper[4873]: E1201 08:40:24.360098 4873 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.362693 4873 server.go:460] "Adding debug handlers to kubelet server" Dec 01 08:40:24 crc kubenswrapper[4873]: E1201 08:40:24.362797 4873 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.195:6443: connect: connection refused" interval="200ms" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.363406 4873 factory.go:153] Registering CRI-O factory Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.363454 4873 factory.go:221] Registration of the crio container factory successfully Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.363696 4873 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.363746 4873 factory.go:103] Registering Raw factory Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.364311 4873 manager.go:1196] Started watching for new ooms in manager Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.368009 4873 manager.go:319] Starting recovery of all containers Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.378099 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.378179 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.378201 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.378223 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.378241 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.378261 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.378279 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.378298 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.378322 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.378340 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.378358 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.378375 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.378393 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.378414 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.378432 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.378450 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.378469 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.378518 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.378536 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.378555 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.378573 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.378593 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.378644 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.378663 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.378680 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.378700 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.378722 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.378771 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.378789 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.378807 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.378823 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.378842 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.378859 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.378876 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.378893 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.378911 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.378926 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.380508 4873 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.380622 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.380643 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.380660 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.380681 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.380699 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.380717 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.380733 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.380750 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.380768 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.380790 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.380811 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.380827 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.380845 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.380861 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.380880 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.380953 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.381377 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.381407 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.381427 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.381447 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.381468 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.381486 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.381504 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.381522 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.381541 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.381562 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.381582 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.381649 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.381671 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.381691 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.381709 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.381727 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.381746 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.381765 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.381781 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.381797 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.381813 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.381830 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.381846 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.381867 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.381884 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.381909 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.381929 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.381946 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.381963 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.381980 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.381996 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.382225 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.382245 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.382263 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.382283 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.382299 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.382315 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.382335 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.382352 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.382369 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.382388 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.382409 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.382425 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.382443 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.382461 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.382479 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.382494 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.382511 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.382528 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.382545 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.382562 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.382588 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.382616 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.382637 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.382656 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.382676 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.382694 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.382720 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.382739 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.382758 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.382777 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.382794 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.382812 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.382830 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.382848 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.382866 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.382883 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.382900 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.382918 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.382936 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.382952 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.382970 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.382987 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383006 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383048 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383068 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383084 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383102 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383118 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383136 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383151 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383168 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383184 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383249 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383266 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383283 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383302 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383321 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383340 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383391 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383407 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383425 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383441 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383458 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383480 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383496 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383512 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383529 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383545 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383558 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383572 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383588 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383604 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383650 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383673 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383695 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383712 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383731 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383750 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383767 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383784 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383800 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383816 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383832 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383849 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383864 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383878 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383894 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383911 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383926 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383942 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383966 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383981 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.383996 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.384011 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.384048 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.384065 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.384081 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.384096 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.384113 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.384126 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.384140 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.384157 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.384174 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.384189 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.384204 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.384223 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.384250 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.384265 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.384286 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.384323 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.384338 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.384353 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.384369 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.384384 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.384402 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.384418 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.384433 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.384448 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.384466 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.384481 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.384497 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.384513 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.384531 4873 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.384546 4873 reconstruct.go:97] "Volume reconstruction finished" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.384573 4873 reconciler.go:26] "Reconciler: start to sync state" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.407421 4873 manager.go:324] Recovery completed Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.424518 4873 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.425474 4873 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.426584 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.426635 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.426647 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.427394 4873 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.427417 4873 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.427442 4873 state_mem.go:36] "Initialized new in-memory state store" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.428489 4873 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.428607 4873 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.428667 4873 kubelet.go:2335] "Starting kubelet main sync loop" Dec 01 08:40:24 crc kubenswrapper[4873]: E1201 08:40:24.428759 4873 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.429935 4873 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.195:6443: connect: connection refused Dec 01 08:40:24 crc kubenswrapper[4873]: E1201 08:40:24.430162 4873 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.195:6443: connect: connection refused" logger="UnhandledError" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.438205 4873 policy_none.go:49] "None policy: Start" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.439696 4873 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.439834 4873 state_mem.go:35] "Initializing new in-memory state store" Dec 01 08:40:24 crc kubenswrapper[4873]: E1201 08:40:24.462984 4873 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.519157 4873 manager.go:334] "Starting Device Plugin manager" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.519990 4873 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.520274 4873 server.go:79] "Starting device plugin registration server" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.521285 4873 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.521492 4873 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.521797 4873 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.521932 4873 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.521947 4873 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.529687 4873 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.529968 4873 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:24 crc kubenswrapper[4873]: E1201 08:40:24.531545 4873 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.531606 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.531952 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.532181 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.532567 4873 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.533002 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.533117 4873 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.534196 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.534240 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.534268 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.534319 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.534350 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.534369 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.534447 4873 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.534840 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.534885 4873 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.535669 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.535695 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.535707 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.535857 4873 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.536093 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.536148 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.536191 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.536242 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.536311 4873 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.537470 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.537520 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.537580 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.537924 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.538002 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.538069 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.538235 4873 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.538345 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.538391 4873 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.542653 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.542734 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.542762 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.544009 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.544067 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.544080 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.544689 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.544783 4873 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.546376 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.546422 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.546440 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:24 crc kubenswrapper[4873]: E1201 08:40:24.564255 4873 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.195:6443: connect: connection refused" interval="400ms" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.587231 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.587337 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.587405 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.587438 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.587501 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.587534 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.587599 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.587632 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.587701 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.587761 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.587794 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.587862 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.587896 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.587962 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.587993 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.622092 4873 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.623858 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.623943 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.623969 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.624009 4873 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 08:40:24 crc kubenswrapper[4873]: E1201 08:40:24.624896 4873 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.195:6443: connect: connection refused" node="crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.689537 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.689679 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.689773 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.689864 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.689970 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.690093 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.690159 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.690191 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.690229 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.690259 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.690288 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.690321 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.690350 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.690344 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.690424 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.690380 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.690474 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.690521 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.690521 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.690599 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.690615 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.690646 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.690675 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.690693 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.690742 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.690786 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.690565 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.690842 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.690912 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.690980 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.825217 4873 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.827518 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.827598 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.827623 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.827669 4873 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 08:40:24 crc kubenswrapper[4873]: E1201 08:40:24.828503 4873 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.195:6443: connect: connection refused" node="crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.865354 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.878463 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.906503 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-18650bb44b7695a0856f9804afa585d2fa070698d692eb6775543e1b394bcfc9 WatchSource:0}: Error finding container 18650bb44b7695a0856f9804afa585d2fa070698d692eb6775543e1b394bcfc9: Status 404 returned error can't find the container with id 18650bb44b7695a0856f9804afa585d2fa070698d692eb6775543e1b394bcfc9 Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.907829 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.909712 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-cdebbe9111614f6a3acaa844f6dda5494a591dc7b810da7817f64e79025d08a0 WatchSource:0}: Error finding container cdebbe9111614f6a3acaa844f6dda5494a591dc7b810da7817f64e79025d08a0: Status 404 returned error can't find the container with id cdebbe9111614f6a3acaa844f6dda5494a591dc7b810da7817f64e79025d08a0 Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.926675 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-05e8f1140e6a8b2e9cb6793eedf84a27611251c4ca9f4888e9bc489fc6ceeb7b WatchSource:0}: Error finding container 05e8f1140e6a8b2e9cb6793eedf84a27611251c4ca9f4888e9bc489fc6ceeb7b: Status 404 returned error can't find the container with id 05e8f1140e6a8b2e9cb6793eedf84a27611251c4ca9f4888e9bc489fc6ceeb7b Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.938289 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: I1201 08:40:24.950766 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:40:24 crc kubenswrapper[4873]: E1201 08:40:24.965311 4873 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.195:6443: connect: connection refused" interval="800ms" Dec 01 08:40:24 crc kubenswrapper[4873]: W1201 08:40:24.967937 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-9b4914ff75379efc47471b9c59d81e49b732e28c8a85e3a417406285a5f6d368 WatchSource:0}: Error finding container 9b4914ff75379efc47471b9c59d81e49b732e28c8a85e3a417406285a5f6d368: Status 404 returned error can't find the container with id 9b4914ff75379efc47471b9c59d81e49b732e28c8a85e3a417406285a5f6d368 Dec 01 08:40:25 crc kubenswrapper[4873]: I1201 08:40:25.229076 4873 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:25 crc kubenswrapper[4873]: I1201 08:40:25.231311 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:25 crc kubenswrapper[4873]: I1201 08:40:25.231376 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:25 crc kubenswrapper[4873]: I1201 08:40:25.231394 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:25 crc kubenswrapper[4873]: I1201 08:40:25.231434 4873 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 08:40:25 crc kubenswrapper[4873]: E1201 08:40:25.232157 4873 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.195:6443: connect: connection refused" node="crc" Dec 01 08:40:25 crc kubenswrapper[4873]: W1201 08:40:25.271327 4873 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.195:6443: connect: connection refused Dec 01 08:40:25 crc kubenswrapper[4873]: E1201 08:40:25.271419 4873 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.195:6443: connect: connection refused" logger="UnhandledError" Dec 01 08:40:25 crc kubenswrapper[4873]: W1201 08:40:25.281871 4873 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.195:6443: connect: connection refused Dec 01 08:40:25 crc kubenswrapper[4873]: E1201 08:40:25.281950 4873 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.195:6443: connect: connection refused" logger="UnhandledError" Dec 01 08:40:25 crc kubenswrapper[4873]: I1201 08:40:25.353214 4873 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.195:6443: connect: connection refused Dec 01 08:40:25 crc kubenswrapper[4873]: W1201 08:40:25.355088 4873 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.195:6443: connect: connection refused Dec 01 08:40:25 crc kubenswrapper[4873]: E1201 08:40:25.355182 4873 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.195:6443: connect: connection refused" logger="UnhandledError" Dec 01 08:40:25 crc kubenswrapper[4873]: I1201 08:40:25.434946 4873 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb" exitCode=0 Dec 01 08:40:25 crc kubenswrapper[4873]: I1201 08:40:25.435029 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb"} Dec 01 08:40:25 crc kubenswrapper[4873]: I1201 08:40:25.435108 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9b4914ff75379efc47471b9c59d81e49b732e28c8a85e3a417406285a5f6d368"} Dec 01 08:40:25 crc kubenswrapper[4873]: I1201 08:40:25.435213 4873 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:25 crc kubenswrapper[4873]: I1201 08:40:25.436082 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:25 crc kubenswrapper[4873]: I1201 08:40:25.436164 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:25 crc kubenswrapper[4873]: I1201 08:40:25.436180 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:25 crc kubenswrapper[4873]: I1201 08:40:25.437323 4873 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268" exitCode=0 Dec 01 08:40:25 crc kubenswrapper[4873]: I1201 08:40:25.437367 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268"} Dec 01 08:40:25 crc kubenswrapper[4873]: I1201 08:40:25.437401 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"05e8f1140e6a8b2e9cb6793eedf84a27611251c4ca9f4888e9bc489fc6ceeb7b"} Dec 01 08:40:25 crc kubenswrapper[4873]: I1201 08:40:25.437540 4873 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:25 crc kubenswrapper[4873]: I1201 08:40:25.437878 4873 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:25 crc kubenswrapper[4873]: I1201 08:40:25.438596 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:25 crc kubenswrapper[4873]: I1201 08:40:25.438627 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:25 crc kubenswrapper[4873]: I1201 08:40:25.438638 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:25 crc kubenswrapper[4873]: I1201 08:40:25.439178 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:25 crc kubenswrapper[4873]: I1201 08:40:25.439204 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:25 crc kubenswrapper[4873]: I1201 08:40:25.439212 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:25 crc kubenswrapper[4873]: I1201 08:40:25.440372 4873 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="24642ae22d2e061a961799d5ca7e07e31897891e2a43cf671bea739030079fab" exitCode=0 Dec 01 08:40:25 crc kubenswrapper[4873]: I1201 08:40:25.440434 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"24642ae22d2e061a961799d5ca7e07e31897891e2a43cf671bea739030079fab"} Dec 01 08:40:25 crc kubenswrapper[4873]: I1201 08:40:25.440455 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"18650bb44b7695a0856f9804afa585d2fa070698d692eb6775543e1b394bcfc9"} Dec 01 08:40:25 crc kubenswrapper[4873]: I1201 08:40:25.440536 4873 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:25 crc kubenswrapper[4873]: I1201 08:40:25.441619 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:25 crc kubenswrapper[4873]: I1201 08:40:25.441647 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:25 crc kubenswrapper[4873]: I1201 08:40:25.441659 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:25 crc kubenswrapper[4873]: I1201 08:40:25.441672 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"aae94f574fbe3c1c7fcd3dbe889f7aa484e1482cee011ebfb2044151d4ca8eac"} Dec 01 08:40:25 crc kubenswrapper[4873]: I1201 08:40:25.441652 4873 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="aae94f574fbe3c1c7fcd3dbe889f7aa484e1482cee011ebfb2044151d4ca8eac" exitCode=0 Dec 01 08:40:25 crc kubenswrapper[4873]: I1201 08:40:25.441771 4873 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:25 crc kubenswrapper[4873]: I1201 08:40:25.441801 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"cdebbe9111614f6a3acaa844f6dda5494a591dc7b810da7817f64e79025d08a0"} Dec 01 08:40:25 crc kubenswrapper[4873]: I1201 08:40:25.442595 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:25 crc kubenswrapper[4873]: I1201 08:40:25.442613 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:25 crc kubenswrapper[4873]: I1201 08:40:25.442622 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:25 crc kubenswrapper[4873]: I1201 08:40:25.443170 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c3ba3cfd8a611c687414d7ecd6b02b423b351d7ac2a4d041085d4ca3bb2b0709"} Dec 01 08:40:25 crc kubenswrapper[4873]: I1201 08:40:25.443190 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"1217ee8d4c06cae59606fa265262699ab70319c9daad7811cf9eec76c2bfb7d8"} Dec 01 08:40:25 crc kubenswrapper[4873]: W1201 08:40:25.729409 4873 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.195:6443: connect: connection refused Dec 01 08:40:25 crc kubenswrapper[4873]: E1201 08:40:25.729592 4873 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.195:6443: connect: connection refused" logger="UnhandledError" Dec 01 08:40:25 crc kubenswrapper[4873]: E1201 08:40:25.766543 4873 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.195:6443: connect: connection refused" interval="1.6s" Dec 01 08:40:26 crc kubenswrapper[4873]: I1201 08:40:26.032520 4873 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:26 crc kubenswrapper[4873]: I1201 08:40:26.034491 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:26 crc kubenswrapper[4873]: I1201 08:40:26.034545 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:26 crc kubenswrapper[4873]: I1201 08:40:26.034556 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:26 crc kubenswrapper[4873]: I1201 08:40:26.034589 4873 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 08:40:26 crc kubenswrapper[4873]: E1201 08:40:26.035409 4873 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.195:6443: connect: connection refused" node="crc" Dec 01 08:40:26 crc kubenswrapper[4873]: I1201 08:40:26.320186 4873 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 01 08:40:26 crc kubenswrapper[4873]: I1201 08:40:26.447123 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568"} Dec 01 08:40:26 crc kubenswrapper[4873]: I1201 08:40:26.447180 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7"} Dec 01 08:40:26 crc kubenswrapper[4873]: I1201 08:40:26.447190 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a"} Dec 01 08:40:26 crc kubenswrapper[4873]: I1201 08:40:26.447200 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9"} Dec 01 08:40:26 crc kubenswrapper[4873]: I1201 08:40:26.448948 4873 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255" exitCode=0 Dec 01 08:40:26 crc kubenswrapper[4873]: I1201 08:40:26.448987 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255"} Dec 01 08:40:26 crc kubenswrapper[4873]: I1201 08:40:26.449101 4873 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:26 crc kubenswrapper[4873]: I1201 08:40:26.449753 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:26 crc kubenswrapper[4873]: I1201 08:40:26.449775 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:26 crc kubenswrapper[4873]: I1201 08:40:26.449786 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:26 crc kubenswrapper[4873]: I1201 08:40:26.452633 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"8cd0a7d74f98d7769eb5513cf477937a5c5f23bc6374335d17281087987d332c"} Dec 01 08:40:26 crc kubenswrapper[4873]: I1201 08:40:26.452828 4873 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:26 crc kubenswrapper[4873]: I1201 08:40:26.454102 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:26 crc kubenswrapper[4873]: I1201 08:40:26.454133 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:26 crc kubenswrapper[4873]: I1201 08:40:26.454144 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:26 crc kubenswrapper[4873]: I1201 08:40:26.455200 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"6ce4df67c597a961222280dff543040c444f4d7eace37c5cfe07c7f9d89ea70a"} Dec 01 08:40:26 crc kubenswrapper[4873]: I1201 08:40:26.455241 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b9ffbc1a66dd17bac0e36d480467c72521d7a97c5376dafe57fa15efe571c78b"} Dec 01 08:40:26 crc kubenswrapper[4873]: I1201 08:40:26.455251 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"25326d274f24d0da05778ed8c5a80d465065df87cf625be2c35d9c95763d4aab"} Dec 01 08:40:26 crc kubenswrapper[4873]: I1201 08:40:26.455258 4873 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:26 crc kubenswrapper[4873]: I1201 08:40:26.457082 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:26 crc kubenswrapper[4873]: I1201 08:40:26.457119 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:26 crc kubenswrapper[4873]: I1201 08:40:26.457130 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:26 crc kubenswrapper[4873]: I1201 08:40:26.466626 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"2c5088b5bf7bc847f59ca9efbf400855a70b99f1bbbf207eb93576c3b3eabae5"} Dec 01 08:40:26 crc kubenswrapper[4873]: I1201 08:40:26.466710 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"1ae6257c73f3632c499b5f107178f64a9cac2f4325c57b41aa75d6d4ea14dd12"} Dec 01 08:40:26 crc kubenswrapper[4873]: I1201 08:40:26.466735 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"8208cd5fda4df2e3b5a8bd90a18c8aca62b16cc2c60e06f52411f24f3d60dd6c"} Dec 01 08:40:26 crc kubenswrapper[4873]: I1201 08:40:26.466874 4873 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:26 crc kubenswrapper[4873]: I1201 08:40:26.469196 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:26 crc kubenswrapper[4873]: I1201 08:40:26.469247 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:26 crc kubenswrapper[4873]: I1201 08:40:26.469261 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:27 crc kubenswrapper[4873]: I1201 08:40:27.474815 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519"} Dec 01 08:40:27 crc kubenswrapper[4873]: I1201 08:40:27.474900 4873 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:27 crc kubenswrapper[4873]: I1201 08:40:27.476569 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:27 crc kubenswrapper[4873]: I1201 08:40:27.476636 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:27 crc kubenswrapper[4873]: I1201 08:40:27.476655 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:27 crc kubenswrapper[4873]: I1201 08:40:27.477714 4873 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9" exitCode=0 Dec 01 08:40:27 crc kubenswrapper[4873]: I1201 08:40:27.477752 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9"} Dec 01 08:40:27 crc kubenswrapper[4873]: I1201 08:40:27.477836 4873 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:27 crc kubenswrapper[4873]: I1201 08:40:27.477922 4873 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:27 crc kubenswrapper[4873]: I1201 08:40:27.477973 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 08:40:27 crc kubenswrapper[4873]: I1201 08:40:27.477934 4873 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:27 crc kubenswrapper[4873]: I1201 08:40:27.479712 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:27 crc kubenswrapper[4873]: I1201 08:40:27.479759 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:27 crc kubenswrapper[4873]: I1201 08:40:27.479814 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:27 crc kubenswrapper[4873]: I1201 08:40:27.479839 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:27 crc kubenswrapper[4873]: I1201 08:40:27.479774 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:27 crc kubenswrapper[4873]: I1201 08:40:27.479929 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:27 crc kubenswrapper[4873]: I1201 08:40:27.480655 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:27 crc kubenswrapper[4873]: I1201 08:40:27.480707 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:27 crc kubenswrapper[4873]: I1201 08:40:27.480728 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:27 crc kubenswrapper[4873]: I1201 08:40:27.635784 4873 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:27 crc kubenswrapper[4873]: I1201 08:40:27.637368 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:27 crc kubenswrapper[4873]: I1201 08:40:27.637427 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:27 crc kubenswrapper[4873]: I1201 08:40:27.637447 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:27 crc kubenswrapper[4873]: I1201 08:40:27.637481 4873 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 08:40:28 crc kubenswrapper[4873]: I1201 08:40:28.437599 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:40:28 crc kubenswrapper[4873]: I1201 08:40:28.445246 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:40:28 crc kubenswrapper[4873]: I1201 08:40:28.486220 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4676d506268107b87ebe2e21ba6e65d959a23a4d6f38bf0de588bd9c5d15e207"} Dec 01 08:40:28 crc kubenswrapper[4873]: I1201 08:40:28.486310 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ff98ccdc9925e44e62f0d6e826ea7926dfa2a7d83d33e475a092f128b7fe920a"} Dec 01 08:40:28 crc kubenswrapper[4873]: I1201 08:40:28.486360 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"7242d1034bd1938cb0bd750cab74db52d454ddd8af75f1e9e0a3c007dcdddd33"} Dec 01 08:40:28 crc kubenswrapper[4873]: I1201 08:40:28.486319 4873 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:28 crc kubenswrapper[4873]: I1201 08:40:28.486401 4873 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:28 crc kubenswrapper[4873]: I1201 08:40:28.486445 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:40:28 crc kubenswrapper[4873]: I1201 08:40:28.486411 4873 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:28 crc kubenswrapper[4873]: I1201 08:40:28.488510 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:28 crc kubenswrapper[4873]: I1201 08:40:28.488578 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:28 crc kubenswrapper[4873]: I1201 08:40:28.488603 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:28 crc kubenswrapper[4873]: I1201 08:40:28.488699 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:28 crc kubenswrapper[4873]: I1201 08:40:28.488730 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:28 crc kubenswrapper[4873]: I1201 08:40:28.488746 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:28 crc kubenswrapper[4873]: I1201 08:40:28.488944 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:28 crc kubenswrapper[4873]: I1201 08:40:28.489092 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:28 crc kubenswrapper[4873]: I1201 08:40:28.489111 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:29 crc kubenswrapper[4873]: I1201 08:40:29.495621 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3acd19dc33ad40d8857d28bc0ccee4c99a6927e92e5029873234ecb5dc6ed18a"} Dec 01 08:40:29 crc kubenswrapper[4873]: I1201 08:40:29.495722 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4751109230b67ad8948a2de4794571581d662bfd8abc6dc5c198dbc411172860"} Dec 01 08:40:29 crc kubenswrapper[4873]: I1201 08:40:29.495771 4873 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:29 crc kubenswrapper[4873]: I1201 08:40:29.495784 4873 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:29 crc kubenswrapper[4873]: I1201 08:40:29.495774 4873 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 08:40:29 crc kubenswrapper[4873]: I1201 08:40:29.495943 4873 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:29 crc kubenswrapper[4873]: I1201 08:40:29.497511 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:29 crc kubenswrapper[4873]: I1201 08:40:29.497559 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:29 crc kubenswrapper[4873]: I1201 08:40:29.497575 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:29 crc kubenswrapper[4873]: I1201 08:40:29.497611 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:29 crc kubenswrapper[4873]: I1201 08:40:29.497650 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:29 crc kubenswrapper[4873]: I1201 08:40:29.497669 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:29 crc kubenswrapper[4873]: I1201 08:40:29.497724 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:29 crc kubenswrapper[4873]: I1201 08:40:29.497757 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:29 crc kubenswrapper[4873]: I1201 08:40:29.497773 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:29 crc kubenswrapper[4873]: I1201 08:40:29.727799 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:40:30 crc kubenswrapper[4873]: I1201 08:40:30.498746 4873 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:30 crc kubenswrapper[4873]: I1201 08:40:30.498804 4873 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:30 crc kubenswrapper[4873]: I1201 08:40:30.500642 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:30 crc kubenswrapper[4873]: I1201 08:40:30.500689 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:30 crc kubenswrapper[4873]: I1201 08:40:30.500706 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:30 crc kubenswrapper[4873]: I1201 08:40:30.500812 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:30 crc kubenswrapper[4873]: I1201 08:40:30.500880 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:30 crc kubenswrapper[4873]: I1201 08:40:30.500901 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:31 crc kubenswrapper[4873]: I1201 08:40:31.353124 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:40:31 crc kubenswrapper[4873]: I1201 08:40:31.502578 4873 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:31 crc kubenswrapper[4873]: I1201 08:40:31.504408 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:31 crc kubenswrapper[4873]: I1201 08:40:31.504484 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:31 crc kubenswrapper[4873]: I1201 08:40:31.504504 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:31 crc kubenswrapper[4873]: I1201 08:40:31.546062 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:40:31 crc kubenswrapper[4873]: I1201 08:40:31.546238 4873 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 08:40:31 crc kubenswrapper[4873]: I1201 08:40:31.546285 4873 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:31 crc kubenswrapper[4873]: I1201 08:40:31.547714 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:31 crc kubenswrapper[4873]: I1201 08:40:31.547785 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:31 crc kubenswrapper[4873]: I1201 08:40:31.547802 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:31 crc kubenswrapper[4873]: I1201 08:40:31.943770 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:40:32 crc kubenswrapper[4873]: I1201 08:40:32.505044 4873 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 08:40:32 crc kubenswrapper[4873]: I1201 08:40:32.505122 4873 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:32 crc kubenswrapper[4873]: I1201 08:40:32.507149 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:32 crc kubenswrapper[4873]: I1201 08:40:32.507230 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:32 crc kubenswrapper[4873]: I1201 08:40:32.507251 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:34 crc kubenswrapper[4873]: I1201 08:40:34.137678 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 01 08:40:34 crc kubenswrapper[4873]: I1201 08:40:34.138321 4873 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:34 crc kubenswrapper[4873]: I1201 08:40:34.140363 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:34 crc kubenswrapper[4873]: I1201 08:40:34.140445 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:34 crc kubenswrapper[4873]: I1201 08:40:34.140467 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:34 crc kubenswrapper[4873]: I1201 08:40:34.275204 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:40:34 crc kubenswrapper[4873]: I1201 08:40:34.275515 4873 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:34 crc kubenswrapper[4873]: I1201 08:40:34.277478 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:34 crc kubenswrapper[4873]: I1201 08:40:34.277564 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:34 crc kubenswrapper[4873]: I1201 08:40:34.277597 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:34 crc kubenswrapper[4873]: E1201 08:40:34.531828 4873 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 01 08:40:34 crc kubenswrapper[4873]: I1201 08:40:34.944135 4873 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 01 08:40:34 crc kubenswrapper[4873]: I1201 08:40:34.944259 4873 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 01 08:40:35 crc kubenswrapper[4873]: I1201 08:40:35.699218 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 01 08:40:35 crc kubenswrapper[4873]: I1201 08:40:35.699448 4873 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:35 crc kubenswrapper[4873]: I1201 08:40:35.700943 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:35 crc kubenswrapper[4873]: I1201 08:40:35.700974 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:35 crc kubenswrapper[4873]: I1201 08:40:35.700985 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:36 crc kubenswrapper[4873]: E1201 08:40:36.321878 4873 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 01 08:40:36 crc kubenswrapper[4873]: I1201 08:40:36.353980 4873 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 01 08:40:36 crc kubenswrapper[4873]: I1201 08:40:36.903562 4873 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 01 08:40:36 crc kubenswrapper[4873]: I1201 08:40:36.904257 4873 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 01 08:40:37 crc kubenswrapper[4873]: I1201 08:40:37.087061 4873 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 01 08:40:37 crc kubenswrapper[4873]: I1201 08:40:37.087126 4873 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 01 08:40:37 crc kubenswrapper[4873]: I1201 08:40:37.092693 4873 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 01 08:40:37 crc kubenswrapper[4873]: I1201 08:40:37.092782 4873 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 01 08:40:40 crc kubenswrapper[4873]: I1201 08:40:40.464114 4873 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 01 08:40:40 crc kubenswrapper[4873]: I1201 08:40:40.487396 4873 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Dec 01 08:40:41 crc kubenswrapper[4873]: I1201 08:40:41.362227 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:40:41 crc kubenswrapper[4873]: I1201 08:40:41.362449 4873 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:41 crc kubenswrapper[4873]: I1201 08:40:41.364120 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:41 crc kubenswrapper[4873]: I1201 08:40:41.364161 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:41 crc kubenswrapper[4873]: I1201 08:40:41.364173 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:41 crc kubenswrapper[4873]: I1201 08:40:41.369747 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:40:41 crc kubenswrapper[4873]: I1201 08:40:41.532112 4873 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:41 crc kubenswrapper[4873]: I1201 08:40:41.534854 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:41 crc kubenswrapper[4873]: I1201 08:40:41.534926 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:41 crc kubenswrapper[4873]: I1201 08:40:41.534951 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:42 crc kubenswrapper[4873]: E1201 08:40:42.088143 4873 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="3.2s" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.090553 4873 trace.go:236] Trace[1913399265]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Dec-2025 08:40:28.232) (total time: 13857ms): Dec 01 08:40:42 crc kubenswrapper[4873]: Trace[1913399265]: ---"Objects listed" error: 13857ms (08:40:42.090) Dec 01 08:40:42 crc kubenswrapper[4873]: Trace[1913399265]: [13.857562127s] [13.857562127s] END Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.090887 4873 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.091744 4873 trace.go:236] Trace[107099179]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Dec-2025 08:40:27.760) (total time: 14331ms): Dec 01 08:40:42 crc kubenswrapper[4873]: Trace[107099179]: ---"Objects listed" error: 14331ms (08:40:42.091) Dec 01 08:40:42 crc kubenswrapper[4873]: Trace[107099179]: [14.331451058s] [14.331451058s] END Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.091781 4873 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.092177 4873 trace.go:236] Trace[1944757457]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Dec-2025 08:40:27.740) (total time: 14351ms): Dec 01 08:40:42 crc kubenswrapper[4873]: Trace[1944757457]: ---"Objects listed" error: 14351ms (08:40:42.091) Dec 01 08:40:42 crc kubenswrapper[4873]: Trace[1944757457]: [14.351258696s] [14.351258696s] END Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.092272 4873 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 01 08:40:42 crc kubenswrapper[4873]: E1201 08:40:42.092512 4873 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.093191 4873 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.093570 4873 trace.go:236] Trace[1989416961]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Dec-2025 08:40:28.732) (total time: 13361ms): Dec 01 08:40:42 crc kubenswrapper[4873]: Trace[1989416961]: ---"Objects listed" error: 13361ms (08:40:42.093) Dec 01 08:40:42 crc kubenswrapper[4873]: Trace[1989416961]: [13.361365882s] [13.361365882s] END Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.093601 4873 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.198952 4873 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:60898->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.198975 4873 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:60890->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.199024 4873 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:60898->192.168.126.11:17697: read: connection reset by peer" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.199076 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:60890->192.168.126.11:17697: read: connection reset by peer" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.199429 4873 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.199514 4873 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.356192 4873 apiserver.go:52] "Watching apiserver" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.359161 4873 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.359517 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.359977 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.360043 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 08:40:42 crc kubenswrapper[4873]: E1201 08:40:42.360073 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.360096 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:40:42 crc kubenswrapper[4873]: E1201 08:40:42.360362 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.360860 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.360898 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.361078 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:40:42 crc kubenswrapper[4873]: E1201 08:40:42.361224 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.363588 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.363598 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.364399 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.364521 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.364600 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.364657 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.364806 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.364901 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.365526 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.392466 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.405319 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.414832 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.425194 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.437212 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.447771 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.457683 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.460426 4873 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.473997 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.495729 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.495814 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.495855 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.495890 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.495934 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.495967 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.496096 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.496140 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.496171 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.496204 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.496238 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.496269 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.496270 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.496303 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.496338 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.496372 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.496370 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.496408 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.496443 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.496477 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.496508 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.496542 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.496576 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.496589 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.496609 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.496648 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.496687 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.496724 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.496765 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.496798 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.496831 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.496866 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.496817 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.496902 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.497091 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.497132 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.497157 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.497182 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.497217 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.497211 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.497273 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.497284 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.497303 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.497385 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.497413 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.497424 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.497446 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.497591 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.497622 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.497613 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.497651 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.497685 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.497688 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.497708 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.497727 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.497742 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.497668 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.497773 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.497756 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.497801 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.497833 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.497858 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.497879 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.497901 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.497926 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.497952 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.497978 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.498007 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.498058 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.498089 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.498119 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.498145 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.498171 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.498177 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.498196 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.498228 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.498261 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.498268 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.498292 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.498322 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.498321 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.498348 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.498353 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.498377 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.498406 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.498426 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.498437 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.498463 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.498460 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.498489 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.498523 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.498548 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.498598 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.498604 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.498625 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.498651 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.498689 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.498720 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.498746 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.498772 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.498796 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.498821 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.498845 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.498871 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.498967 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.498997 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.499038 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.499067 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.499095 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.499139 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.499187 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.499192 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.499192 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.499263 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.499294 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.499323 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.499334 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.499454 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.499713 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.500219 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.500245 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.500403 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.500425 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.500506 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.500525 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.500496 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.500569 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.500810 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.501344 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.501394 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.501749 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.501773 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.501816 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.502076 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.502315 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.499369 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.502726 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.502765 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.502791 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.502801 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.502944 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.502984 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.502988 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.503035 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.503065 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.503095 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.503125 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.503170 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.503199 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.503224 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.503253 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.503281 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.503308 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.503331 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.503369 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.503396 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.503420 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.503449 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.503472 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.503494 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.503518 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.503547 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.503574 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.503596 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.503620 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.503645 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.503733 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.503799 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.503855 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.503890 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.503916 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.503943 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.503978 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.504007 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.504058 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.504083 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.504111 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.504138 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.504164 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.504197 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.504222 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.504314 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.504343 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.504368 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.504394 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.504418 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.504450 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.504475 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.504499 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.504525 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.504550 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.504574 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.504601 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.504625 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.504650 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.504676 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.504703 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.504728 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.504754 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.504779 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.504823 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.504850 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.504874 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.504906 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.504933 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.504956 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.504981 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.505077 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.505150 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.505180 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.505204 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.505232 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.505258 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.505285 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.505309 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.505336 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.505357 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.505380 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.505403 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.505426 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.505456 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.505482 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.505507 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.505531 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.505562 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.505591 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.505618 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.505645 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.505681 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.505708 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.505731 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.505756 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.505786 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.505810 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.505834 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.505860 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.505895 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.505922 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.505947 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.505977 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506002 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506082 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506117 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506146 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506175 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506205 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506236 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506271 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506297 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506323 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506350 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506374 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506405 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506434 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506461 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506579 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506601 4873 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506617 4873 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506632 4873 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506646 4873 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506660 4873 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506672 4873 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506687 4873 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506701 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506717 4873 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506733 4873 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506745 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506758 4873 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506769 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506781 4873 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506793 4873 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506805 4873 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506819 4873 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506832 4873 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506848 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506861 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506878 4873 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506891 4873 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506904 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506916 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506929 4873 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506941 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506953 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506965 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506977 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506990 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.507001 4873 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.507031 4873 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.507042 4873 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.507053 4873 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.507064 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.507077 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.507090 4873 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.507102 4873 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.507118 4873 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.507129 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.507143 4873 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.507154 4873 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.507165 4873 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.507178 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.507190 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.503975 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.504072 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.504082 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.504167 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.504295 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.507794 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.504732 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.504874 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.504884 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.505246 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.505526 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.505558 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.505664 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.505890 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506062 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506111 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506189 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506508 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506523 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506589 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506797 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506848 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.506952 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.507122 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.507205 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.507284 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.507346 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.507452 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.508167 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.507628 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.508273 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.507945 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.508345 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.508426 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.508568 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.508723 4873 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.508856 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.508961 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.509246 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.509398 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.510133 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.510352 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.510586 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.510612 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.511564 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.511591 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.511907 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.511971 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.512401 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.512554 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.512836 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.512916 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.512987 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.513034 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.513201 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.513359 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.513523 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.513588 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.513900 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.514401 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.514400 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.514521 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.514409 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.514449 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.514577 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.514858 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.514917 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.514949 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.514967 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: E1201 08:40:42.515541 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:40:43.015508699 +0000 UTC m=+18.917617398 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.515976 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.516071 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.516223 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: E1201 08:40:42.516519 4873 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.516963 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: E1201 08:40:42.517080 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:40:43.017040039 +0000 UTC m=+18.919148688 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:40:42 crc kubenswrapper[4873]: E1201 08:40:42.516554 4873 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:40:42 crc kubenswrapper[4873]: E1201 08:40:42.517174 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:40:43.017152852 +0000 UTC m=+18.919261471 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.516825 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.517482 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.517606 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.518115 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.518212 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.518464 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.520472 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.521255 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.516973 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.521396 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.521427 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.517856 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.525145 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.525346 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.525470 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.525870 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.525948 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: E1201 08:40:42.526076 4873 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:40:42 crc kubenswrapper[4873]: E1201 08:40:42.526101 4873 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:40:42 crc kubenswrapper[4873]: E1201 08:40:42.526121 4873 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.526144 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: E1201 08:40:42.526220 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 08:40:43.026192569 +0000 UTC m=+18.928301298 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:42 crc kubenswrapper[4873]: E1201 08:40:42.526942 4873 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:40:42 crc kubenswrapper[4873]: E1201 08:40:42.526975 4873 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:40:42 crc kubenswrapper[4873]: E1201 08:40:42.526995 4873 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:42 crc kubenswrapper[4873]: E1201 08:40:42.527341 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 08:40:43.027313948 +0000 UTC m=+18.929422657 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.527689 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.527693 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.528439 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.528503 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.528793 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.529304 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.529380 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.529509 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.529604 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.529712 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.530300 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.530357 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.530568 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.530940 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.531227 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.531370 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.531398 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.532097 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.532752 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.533521 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.534142 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.534596 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.534852 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.535806 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.535854 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.536185 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.536300 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.536419 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.536421 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.536509 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.536921 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.537126 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.537306 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.537437 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.537518 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.537563 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.537589 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.537537 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.538062 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.539009 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.539195 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.539223 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.539393 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.542517 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.542675 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.543711 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.543864 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.544489 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.544488 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.544932 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.544994 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.545139 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.545475 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.545637 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.545824 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.545824 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.546570 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.546796 4873 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519" exitCode=255 Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.546856 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519"} Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.547115 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.547246 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.547877 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.547969 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.547972 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.548941 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.560683 4873 scope.go:117] "RemoveContainer" containerID="f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.561089 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.565269 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.566387 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.571367 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.573504 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.580145 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.580253 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.583122 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.584168 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.587684 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.594278 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.605846 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.609171 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.609222 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.609270 4873 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.609281 4873 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.609292 4873 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.609300 4873 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.609309 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.609317 4873 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.609326 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.609333 4873 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.609341 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.609375 4873 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.609383 4873 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.609494 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.609467 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.609616 4873 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.609640 4873 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.609664 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.609684 4873 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.609836 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.609850 4873 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.609862 4873 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.609874 4873 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.609886 4873 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.609898 4873 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.609909 4873 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.609923 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.609935 4873 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.609946 4873 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.609958 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.609971 4873 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.609981 4873 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.609991 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610005 4873 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610039 4873 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610052 4873 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610064 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610078 4873 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610283 4873 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610298 4873 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610310 4873 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610322 4873 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610334 4873 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610352 4873 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610365 4873 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610377 4873 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610389 4873 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610401 4873 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610413 4873 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610425 4873 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610439 4873 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610452 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610475 4873 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610488 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610500 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610516 4873 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610529 4873 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610545 4873 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610558 4873 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610571 4873 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610585 4873 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610597 4873 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610611 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610624 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610636 4873 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610649 4873 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610660 4873 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610673 4873 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610685 4873 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610698 4873 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610712 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610727 4873 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610739 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610752 4873 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610764 4873 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610775 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610789 4873 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610802 4873 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610814 4873 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610826 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610840 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610852 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610865 4873 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610876 4873 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610888 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610902 4873 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610913 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610924 4873 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610937 4873 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610950 4873 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610962 4873 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610978 4873 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.610992 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611007 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611035 4873 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611051 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611066 4873 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611079 4873 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611092 4873 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611104 4873 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611117 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611132 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611145 4873 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611158 4873 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611172 4873 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611208 4873 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611223 4873 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611235 4873 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611248 4873 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611259 4873 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611271 4873 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611283 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611296 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611309 4873 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611323 4873 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611334 4873 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611347 4873 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611361 4873 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611372 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611385 4873 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611397 4873 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611410 4873 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611422 4873 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611435 4873 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611447 4873 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611460 4873 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611471 4873 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611484 4873 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611497 4873 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611509 4873 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611525 4873 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611538 4873 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611551 4873 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611566 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611578 4873 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611591 4873 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611604 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611640 4873 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611655 4873 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611668 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611681 4873 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611693 4873 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611706 4873 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611719 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611733 4873 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611746 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611760 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611773 4873 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611785 4873 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611799 4873 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611823 4873 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.611836 4873 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.616219 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.629236 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.642912 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.652867 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.655857 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.663095 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.672928 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.674921 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.684114 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1c8f149-f188-4759-b2dc-9dc15be1bb13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:40:37.024907 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:40:37.026665 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1609525021/tls.crt::/tmp/serving-cert-1609525021/tls.key\\\\\\\"\\\\nI1201 08:40:42.169070 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:40:42.177896 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:40:42.177940 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:40:42.177968 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:40:42.177975 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:40:42.184767 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1201 08:40:42.184781 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1201 08:40:42.184802 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184808 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184814 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:40:42.184819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:40:42.184822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:40:42.184826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1201 08:40:42.187129 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.685096 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 08:40:42 crc kubenswrapper[4873]: W1201 08:40:42.686445 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-50b7731e6ba61df3bad31146e26536fdf0e02d0e9004ab676ce374ea578bac27 WatchSource:0}: Error finding container 50b7731e6ba61df3bad31146e26536fdf0e02d0e9004ab676ce374ea578bac27: Status 404 returned error can't find the container with id 50b7731e6ba61df3bad31146e26536fdf0e02d0e9004ab676ce374ea578bac27 Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.693881 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.696217 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:40:42 crc kubenswrapper[4873]: I1201 08:40:42.705343 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:40:43 crc kubenswrapper[4873]: I1201 08:40:43.115989 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:40:43 crc kubenswrapper[4873]: I1201 08:40:43.116160 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:40:43 crc kubenswrapper[4873]: I1201 08:40:43.116191 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:40:43 crc kubenswrapper[4873]: E1201 08:40:43.116244 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:40:44.116196558 +0000 UTC m=+20.018305097 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:40:43 crc kubenswrapper[4873]: I1201 08:40:43.116302 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:40:43 crc kubenswrapper[4873]: I1201 08:40:43.116363 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:40:43 crc kubenswrapper[4873]: E1201 08:40:43.116314 4873 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:40:43 crc kubenswrapper[4873]: E1201 08:40:43.116511 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:40:44.116490636 +0000 UTC m=+20.018599185 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:40:43 crc kubenswrapper[4873]: E1201 08:40:43.116518 4873 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:40:43 crc kubenswrapper[4873]: E1201 08:40:43.116538 4873 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:40:43 crc kubenswrapper[4873]: E1201 08:40:43.116550 4873 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:43 crc kubenswrapper[4873]: E1201 08:40:43.116595 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 08:40:44.116584178 +0000 UTC m=+20.018692767 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:43 crc kubenswrapper[4873]: E1201 08:40:43.116347 4873 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:40:43 crc kubenswrapper[4873]: E1201 08:40:43.116637 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:40:44.11663005 +0000 UTC m=+20.018738689 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:40:43 crc kubenswrapper[4873]: E1201 08:40:43.116438 4873 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:40:43 crc kubenswrapper[4873]: E1201 08:40:43.116656 4873 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:40:43 crc kubenswrapper[4873]: E1201 08:40:43.116665 4873 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:43 crc kubenswrapper[4873]: E1201 08:40:43.116691 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 08:40:44.116683761 +0000 UTC m=+20.018792390 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:43 crc kubenswrapper[4873]: I1201 08:40:43.224215 4873 csr.go:261] certificate signing request csr-rkwfx is approved, waiting to be issued Dec 01 08:40:43 crc kubenswrapper[4873]: I1201 08:40:43.249575 4873 csr.go:257] certificate signing request csr-rkwfx is issued Dec 01 08:40:43 crc kubenswrapper[4873]: I1201 08:40:43.550557 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"bf1fac1203f282ede8c800eef6e0e74d88150766cda6a8863c683bfa16220c39"} Dec 01 08:40:43 crc kubenswrapper[4873]: I1201 08:40:43.552132 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"90f08b6fa658ce25fbc25d8d44d956cdb6424a0b047205f19d1a6fd6870b1e90"} Dec 01 08:40:43 crc kubenswrapper[4873]: I1201 08:40:43.552162 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"9a8d27e3678f1f63f55bd1b76668e328f49368d67d4d7338a6326cb73d13e08f"} Dec 01 08:40:43 crc kubenswrapper[4873]: I1201 08:40:43.552175 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"c44f74df18ef1d07195ae0c25ab7279e7a6ab64b8f77991bfd7c0818d163530e"} Dec 01 08:40:43 crc kubenswrapper[4873]: I1201 08:40:43.554088 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"88068a1cf89e2c507b9d93d305db6af5af196bd9eec805f42d91b6fe090aba35"} Dec 01 08:40:43 crc kubenswrapper[4873]: I1201 08:40:43.554117 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"50b7731e6ba61df3bad31146e26536fdf0e02d0e9004ab676ce374ea578bac27"} Dec 01 08:40:43 crc kubenswrapper[4873]: I1201 08:40:43.556395 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 01 08:40:43 crc kubenswrapper[4873]: I1201 08:40:43.558868 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9"} Dec 01 08:40:43 crc kubenswrapper[4873]: I1201 08:40:43.569207 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4873]: I1201 08:40:43.590240 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4873]: I1201 08:40:43.605586 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90f08b6fa658ce25fbc25d8d44d956cdb6424a0b047205f19d1a6fd6870b1e90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8d27e3678f1f63f55bd1b76668e328f49368d67d4d7338a6326cb73d13e08f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4873]: I1201 08:40:43.616689 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4873]: I1201 08:40:43.635708 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1c8f149-f188-4759-b2dc-9dc15be1bb13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:40:37.024907 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:40:37.026665 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1609525021/tls.crt::/tmp/serving-cert-1609525021/tls.key\\\\\\\"\\\\nI1201 08:40:42.169070 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:40:42.177896 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:40:42.177940 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:40:42.177968 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:40:42.177975 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:40:42.184767 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1201 08:40:42.184781 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1201 08:40:42.184802 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184808 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184814 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:40:42.184819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:40:42.184822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:40:42.184826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1201 08:40:42.187129 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4873]: I1201 08:40:43.651146 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f7ac0b4-d048-46e4-bcee-4a271503e46d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8208cd5fda4df2e3b5a8bd90a18c8aca62b16cc2c60e06f52411f24f3d60dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba3cfd8a611c687414d7ecd6b02b423b351d7ac2a4d041085d4ca3bb2b0709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ae6257c73f3632c499b5f107178f64a9cac2f4325c57b41aa75d6d4ea14dd12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5088b5bf7bc847f59ca9efbf400855a70b99f1bbbf207eb93576c3b3eabae5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4873]: I1201 08:40:43.667267 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4873]: I1201 08:40:43.680560 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4873]: I1201 08:40:43.694285 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88068a1cf89e2c507b9d93d305db6af5af196bd9eec805f42d91b6fe090aba35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4873]: I1201 08:40:43.717303 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4873]: I1201 08:40:43.745986 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90f08b6fa658ce25fbc25d8d44d956cdb6424a0b047205f19d1a6fd6870b1e90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8d27e3678f1f63f55bd1b76668e328f49368d67d4d7338a6326cb73d13e08f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4873]: I1201 08:40:43.775223 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4873]: I1201 08:40:43.815076 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4873]: I1201 08:40:43.834123 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1c8f149-f188-4759-b2dc-9dc15be1bb13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:40:37.024907 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:40:37.026665 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1609525021/tls.crt::/tmp/serving-cert-1609525021/tls.key\\\\\\\"\\\\nI1201 08:40:42.169070 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:40:42.177896 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:40:42.177940 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:40:42.177968 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:40:42.177975 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:40:42.184767 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1201 08:40:42.184781 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1201 08:40:42.184802 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184808 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184814 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:40:42.184819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:40:42.184822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:40:42.184826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1201 08:40:42.187129 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4873]: I1201 08:40:43.860260 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f7ac0b4-d048-46e4-bcee-4a271503e46d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8208cd5fda4df2e3b5a8bd90a18c8aca62b16cc2c60e06f52411f24f3d60dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba3cfd8a611c687414d7ecd6b02b423b351d7ac2a4d041085d4ca3bb2b0709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ae6257c73f3632c499b5f107178f64a9cac2f4325c57b41aa75d6d4ea14dd12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5088b5bf7bc847f59ca9efbf400855a70b99f1bbbf207eb93576c3b3eabae5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4873]: I1201 08:40:43.879638 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4873]: I1201 08:40:43.992266 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-scwpp"] Dec 01 08:40:43 crc kubenswrapper[4873]: I1201 08:40:43.992709 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" Dec 01 08:40:43 crc kubenswrapper[4873]: I1201 08:40:43.994194 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-qpr6r"] Dec 01 08:40:43 crc kubenswrapper[4873]: I1201 08:40:43.994411 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-qpr6r" Dec 01 08:40:43 crc kubenswrapper[4873]: I1201 08:40:43.996221 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 01 08:40:43 crc kubenswrapper[4873]: I1201 08:40:43.996289 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 01 08:40:43 crc kubenswrapper[4873]: I1201 08:40:43.996422 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 01 08:40:43 crc kubenswrapper[4873]: I1201 08:40:43.996476 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 01 08:40:43 crc kubenswrapper[4873]: I1201 08:40:43.996567 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 01 08:40:43 crc kubenswrapper[4873]: I1201 08:40:43.996840 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 01 08:40:43 crc kubenswrapper[4873]: I1201 08:40:43.997498 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 01 08:40:43 crc kubenswrapper[4873]: I1201 08:40:43.998641 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.021295 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1c8f149-f188-4759-b2dc-9dc15be1bb13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:40:37.024907 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:40:37.026665 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1609525021/tls.crt::/tmp/serving-cert-1609525021/tls.key\\\\\\\"\\\\nI1201 08:40:42.169070 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:40:42.177896 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:40:42.177940 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:40:42.177968 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:40:42.177975 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:40:42.184767 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1201 08:40:42.184781 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1201 08:40:42.184802 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184808 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184814 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:40:42.184819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:40:42.184822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:40:42.184826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1201 08:40:42.187129 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.038432 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f7ac0b4-d048-46e4-bcee-4a271503e46d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8208cd5fda4df2e3b5a8bd90a18c8aca62b16cc2c60e06f52411f24f3d60dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba3cfd8a611c687414d7ecd6b02b423b351d7ac2a4d041085d4ca3bb2b0709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ae6257c73f3632c499b5f107178f64a9cac2f4325c57b41aa75d6d4ea14dd12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5088b5bf7bc847f59ca9efbf400855a70b99f1bbbf207eb93576c3b3eabae5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.080330 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.101352 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.123567 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.123666 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/fef7b114-0e07-402d-a37b-315c36011f4b-rootfs\") pod \"machine-config-daemon-scwpp\" (UID: \"fef7b114-0e07-402d-a37b-315c36011f4b\") " pod="openshift-machine-config-operator/machine-config-daemon-scwpp" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.123693 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bqq5\" (UniqueName: \"kubernetes.io/projected/bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0-kube-api-access-6bqq5\") pod \"node-resolver-qpr6r\" (UID: \"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0\") " pod="openshift-dns/node-resolver-qpr6r" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.123719 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.123738 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fef7b114-0e07-402d-a37b-315c36011f4b-mcd-auth-proxy-config\") pod \"machine-config-daemon-scwpp\" (UID: \"fef7b114-0e07-402d-a37b-315c36011f4b\") " pod="openshift-machine-config-operator/machine-config-daemon-scwpp" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.123758 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgsbx\" (UniqueName: \"kubernetes.io/projected/fef7b114-0e07-402d-a37b-315c36011f4b-kube-api-access-mgsbx\") pod \"machine-config-daemon-scwpp\" (UID: \"fef7b114-0e07-402d-a37b-315c36011f4b\") " pod="openshift-machine-config-operator/machine-config-daemon-scwpp" Dec 01 08:40:44 crc kubenswrapper[4873]: E1201 08:40:44.123803 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:40:46.123767348 +0000 UTC m=+22.025875887 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.123868 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fef7b114-0e07-402d-a37b-315c36011f4b-proxy-tls\") pod \"machine-config-daemon-scwpp\" (UID: \"fef7b114-0e07-402d-a37b-315c36011f4b\") " pod="openshift-machine-config-operator/machine-config-daemon-scwpp" Dec 01 08:40:44 crc kubenswrapper[4873]: E1201 08:40:44.123915 4873 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:40:44 crc kubenswrapper[4873]: E1201 08:40:44.123933 4873 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:40:44 crc kubenswrapper[4873]: E1201 08:40:44.123947 4873 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:44 crc kubenswrapper[4873]: E1201 08:40:44.123994 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 08:40:46.123976573 +0000 UTC m=+22.026085112 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:44 crc kubenswrapper[4873]: E1201 08:40:44.124078 4873 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:40:44 crc kubenswrapper[4873]: E1201 08:40:44.124132 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:40:46.124124467 +0000 UTC m=+22.026233006 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.123928 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.124195 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.124219 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0-hosts-file\") pod \"node-resolver-qpr6r\" (UID: \"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0\") " pod="openshift-dns/node-resolver-qpr6r" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.124237 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:40:44 crc kubenswrapper[4873]: E1201 08:40:44.124316 4873 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:40:44 crc kubenswrapper[4873]: E1201 08:40:44.124330 4873 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:40:44 crc kubenswrapper[4873]: E1201 08:40:44.124339 4873 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:40:44 crc kubenswrapper[4873]: E1201 08:40:44.124341 4873 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:44 crc kubenswrapper[4873]: E1201 08:40:44.124375 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:40:46.124368563 +0000 UTC m=+22.026477102 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:40:44 crc kubenswrapper[4873]: E1201 08:40:44.124392 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 08:40:46.124384914 +0000 UTC m=+22.026493453 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.132813 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef7b114-0e07-402d-a37b-315c36011f4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-scwpp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.151619 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.186439 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90f08b6fa658ce25fbc25d8d44d956cdb6424a0b047205f19d1a6fd6870b1e90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8d27e3678f1f63f55bd1b76668e328f49368d67d4d7338a6326cb73d13e08f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.198446 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.211595 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88068a1cf89e2c507b9d93d305db6af5af196bd9eec805f42d91b6fe090aba35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.225259 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0-hosts-file\") pod \"node-resolver-qpr6r\" (UID: \"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0\") " pod="openshift-dns/node-resolver-qpr6r" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.225288 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.225351 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/fef7b114-0e07-402d-a37b-315c36011f4b-rootfs\") pod \"machine-config-daemon-scwpp\" (UID: \"fef7b114-0e07-402d-a37b-315c36011f4b\") " pod="openshift-machine-config-operator/machine-config-daemon-scwpp" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.225307 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/fef7b114-0e07-402d-a37b-315c36011f4b-rootfs\") pod \"machine-config-daemon-scwpp\" (UID: \"fef7b114-0e07-402d-a37b-315c36011f4b\") " pod="openshift-machine-config-operator/machine-config-daemon-scwpp" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.225405 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0-hosts-file\") pod \"node-resolver-qpr6r\" (UID: \"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0\") " pod="openshift-dns/node-resolver-qpr6r" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.225418 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bqq5\" (UniqueName: \"kubernetes.io/projected/bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0-kube-api-access-6bqq5\") pod \"node-resolver-qpr6r\" (UID: \"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0\") " pod="openshift-dns/node-resolver-qpr6r" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.225438 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fef7b114-0e07-402d-a37b-315c36011f4b-mcd-auth-proxy-config\") pod \"machine-config-daemon-scwpp\" (UID: \"fef7b114-0e07-402d-a37b-315c36011f4b\") " pod="openshift-machine-config-operator/machine-config-daemon-scwpp" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.225460 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgsbx\" (UniqueName: \"kubernetes.io/projected/fef7b114-0e07-402d-a37b-315c36011f4b-kube-api-access-mgsbx\") pod \"machine-config-daemon-scwpp\" (UID: \"fef7b114-0e07-402d-a37b-315c36011f4b\") " pod="openshift-machine-config-operator/machine-config-daemon-scwpp" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.225476 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fef7b114-0e07-402d-a37b-315c36011f4b-proxy-tls\") pod \"machine-config-daemon-scwpp\" (UID: \"fef7b114-0e07-402d-a37b-315c36011f4b\") " pod="openshift-machine-config-operator/machine-config-daemon-scwpp" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.226087 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fef7b114-0e07-402d-a37b-315c36011f4b-mcd-auth-proxy-config\") pod \"machine-config-daemon-scwpp\" (UID: \"fef7b114-0e07-402d-a37b-315c36011f4b\") " pod="openshift-machine-config-operator/machine-config-daemon-scwpp" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.237034 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fef7b114-0e07-402d-a37b-315c36011f4b-proxy-tls\") pod \"machine-config-daemon-scwpp\" (UID: \"fef7b114-0e07-402d-a37b-315c36011f4b\") " pod="openshift-machine-config-operator/machine-config-daemon-scwpp" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.240818 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.242571 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bqq5\" (UniqueName: \"kubernetes.io/projected/bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0-kube-api-access-6bqq5\") pod \"node-resolver-qpr6r\" (UID: \"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0\") " pod="openshift-dns/node-resolver-qpr6r" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.244496 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgsbx\" (UniqueName: \"kubernetes.io/projected/fef7b114-0e07-402d-a37b-315c36011f4b-kube-api-access-mgsbx\") pod \"machine-config-daemon-scwpp\" (UID: \"fef7b114-0e07-402d-a37b-315c36011f4b\") " pod="openshift-machine-config-operator/machine-config-daemon-scwpp" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.249241 4873 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Dec 01 08:40:44 crc kubenswrapper[4873]: W1201 08:40:44.250049 4873 reflector.go:484] object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq": watch of *v1.Secret ended with: very short watch: object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq": Unexpected watch close - watch lasted less than a second and no items received Dec 01 08:40:44 crc kubenswrapper[4873]: W1201 08:40:44.250085 4873 reflector.go:484] object-"openshift-machine-config-operator"/"kube-rbac-proxy": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-machine-config-operator"/"kube-rbac-proxy": Unexpected watch close - watch lasted less than a second and no items received Dec 01 08:40:44 crc kubenswrapper[4873]: W1201 08:40:44.250384 4873 reflector.go:484] object-"openshift-dns"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-dns"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 01 08:40:44 crc kubenswrapper[4873]: W1201 08:40:44.250442 4873 reflector.go:484] object-"openshift-dns"/"node-resolver-dockercfg-kz9s7": watch of *v1.Secret ended with: very short watch: object-"openshift-dns"/"node-resolver-dockercfg-kz9s7": Unexpected watch close - watch lasted less than a second and no items received Dec 01 08:40:44 crc kubenswrapper[4873]: W1201 08:40:44.250694 4873 reflector.go:484] object-"openshift-dns"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-dns"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 01 08:40:44 crc kubenswrapper[4873]: W1201 08:40:44.250732 4873 reflector.go:484] object-"openshift-machine-config-operator"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-machine-config-operator"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 01 08:40:44 crc kubenswrapper[4873]: W1201 08:40:44.250759 4873 reflector.go:484] object-"openshift-machine-config-operator"/"proxy-tls": watch of *v1.Secret ended with: very short watch: object-"openshift-machine-config-operator"/"proxy-tls": Unexpected watch close - watch lasted less than a second and no items received Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.250863 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef7b114-0e07-402d-a37b-315c36011f4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-scwpp\": Patch \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-scwpp/status\": read tcp 38.102.83.195:44452->38.102.83.195:6443: use of closed network connection" Dec 01 08:40:44 crc kubenswrapper[4873]: W1201 08:40:44.251229 4873 reflector.go:484] object-"openshift-machine-config-operator"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-machine-config-operator"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.251303 4873 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-12-01 08:35:43 +0000 UTC, rotation deadline is 2026-10-09 00:10:41.897559605 +0000 UTC Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.251348 4873 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 7479h29m57.646214874s for next certificate rotation Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.279693 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1c8f149-f188-4759-b2dc-9dc15be1bb13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:40:37.024907 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:40:37.026665 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1609525021/tls.crt::/tmp/serving-cert-1609525021/tls.key\\\\\\\"\\\\nI1201 08:40:42.169070 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:40:42.177896 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:40:42.177940 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:40:42.177968 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:40:42.177975 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:40:42.184767 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1201 08:40:42.184781 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1201 08:40:42.184802 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184808 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184814 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:40:42.184819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:40:42.184822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:40:42.184826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1201 08:40:42.187129 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.292220 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f7ac0b4-d048-46e4-bcee-4a271503e46d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8208cd5fda4df2e3b5a8bd90a18c8aca62b16cc2c60e06f52411f24f3d60dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba3cfd8a611c687414d7ecd6b02b423b351d7ac2a4d041085d4ca3bb2b0709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ae6257c73f3632c499b5f107178f64a9cac2f4325c57b41aa75d6d4ea14dd12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5088b5bf7bc847f59ca9efbf400855a70b99f1bbbf207eb93576c3b3eabae5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.306536 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.313505 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.318230 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-qpr6r" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.324709 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qpr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qpr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: W1201 08:40:44.335296 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbcf3f8dc_c2aa_4e0a_a842_e053dd5bc0b0.slice/crio-3b8f12cd789caf08587b20ffe02edf137dae3e713b7f70d2c36e5af9e65f2dba WatchSource:0}: Error finding container 3b8f12cd789caf08587b20ffe02edf137dae3e713b7f70d2c36e5af9e65f2dba: Status 404 returned error can't find the container with id 3b8f12cd789caf08587b20ffe02edf137dae3e713b7f70d2c36e5af9e65f2dba Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.342786 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88068a1cf89e2c507b9d93d305db6af5af196bd9eec805f42d91b6fe090aba35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.364354 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.380796 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90f08b6fa658ce25fbc25d8d44d956cdb6424a0b047205f19d1a6fd6870b1e90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8d27e3678f1f63f55bd1b76668e328f49368d67d4d7338a6326cb73d13e08f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.392384 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-shkn9"] Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.392999 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-shkn9" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.393122 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-dxlvk"] Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.393657 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-tpwwr"] Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.393813 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.394055 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.396419 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.400109 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.400231 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.400723 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 01 08:40:44 crc kubenswrapper[4873]: W1201 08:40:44.400845 4873 reflector.go:561] object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl": failed to list *v1.Secret: secrets "ovn-kubernetes-node-dockercfg-pwtwl" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'crc' and this object Dec 01 08:40:44 crc kubenswrapper[4873]: E1201 08:40:44.400878 4873 reflector.go:158] "Unhandled Error" err="object-\"openshift-ovn-kubernetes\"/\"ovn-kubernetes-node-dockercfg-pwtwl\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"ovn-kubernetes-node-dockercfg-pwtwl\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-ovn-kubernetes\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.400933 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.401181 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.401288 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.401354 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.402618 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.402655 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.403148 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.403615 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.405470 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.413412 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1c8f149-f188-4759-b2dc-9dc15be1bb13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:40:37.024907 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:40:37.026665 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1609525021/tls.crt::/tmp/serving-cert-1609525021/tls.key\\\\\\\"\\\\nI1201 08:40:42.169070 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:40:42.177896 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:40:42.177940 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:40:42.177968 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:40:42.177975 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:40:42.184767 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1201 08:40:42.184781 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1201 08:40:42.184802 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184808 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184814 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:40:42.184819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:40:42.184822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:40:42.184826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1201 08:40:42.187129 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.430677 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:40:44 crc kubenswrapper[4873]: E1201 08:40:44.430844 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.430929 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:40:44 crc kubenswrapper[4873]: E1201 08:40:44.431143 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.431276 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:40:44 crc kubenswrapper[4873]: E1201 08:40:44.431463 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.432660 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.433522 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.436869 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.437770 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.438896 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.439451 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.440091 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.441198 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.441458 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90f08b6fa658ce25fbc25d8d44d956cdb6424a0b047205f19d1a6fd6870b1e90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8d27e3678f1f63f55bd1b76668e328f49368d67d4d7338a6326cb73d13e08f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.441871 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.442892 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.443424 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.444616 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.445168 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.445670 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.446612 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.447167 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.448124 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.448556 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.449114 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.450217 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.450710 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.451809 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.452272 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.453967 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.454478 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.455155 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.455452 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qpr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qpr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.456310 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.456791 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.458326 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.459552 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.460619 4873 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.460727 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.462499 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.464090 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.464645 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.467171 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.472145 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.472949 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.473650 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.475007 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.475551 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.475457 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8570b876-a13b-4b8f-99c4-4900cfe19a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shkn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.478578 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.483145 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.483895 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.484895 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.485490 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.486778 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.487723 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.488879 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.489536 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.490124 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.491268 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.491940 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.492972 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.493094 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f7ac0b4-d048-46e4-bcee-4a271503e46d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8208cd5fda4df2e3b5a8bd90a18c8aca62b16cc2c60e06f52411f24f3d60dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba3cfd8a611c687414d7ecd6b02b423b351d7ac2a4d041085d4ca3bb2b0709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ae6257c73f3632c499b5f107178f64a9cac2f4325c57b41aa75d6d4ea14dd12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5088b5bf7bc847f59ca9efbf400855a70b99f1bbbf207eb93576c3b3eabae5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.517062 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.531974 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-run-ovn\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.532037 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/13a2246b-93bb-4586-98ee-53fc84aaae02-host-run-multus-certs\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.532057 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/13a2246b-93bb-4586-98ee-53fc84aaae02-host-var-lib-kubelet\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.532080 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxfd8\" (UniqueName: \"kubernetes.io/projected/a308045b-f54d-42ec-8fdf-5bc3e54ef363-kube-api-access-dxfd8\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.532105 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8570b876-a13b-4b8f-99c4-4900cfe19a21-system-cni-dir\") pod \"multus-additional-cni-plugins-shkn9\" (UID: \"8570b876-a13b-4b8f-99c4-4900cfe19a21\") " pod="openshift-multus/multus-additional-cni-plugins-shkn9" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.532135 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbkv6\" (UniqueName: \"kubernetes.io/projected/8570b876-a13b-4b8f-99c4-4900cfe19a21-kube-api-access-mbkv6\") pod \"multus-additional-cni-plugins-shkn9\" (UID: \"8570b876-a13b-4b8f-99c4-4900cfe19a21\") " pod="openshift-multus/multus-additional-cni-plugins-shkn9" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.532153 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-host-run-netns\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.532175 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/8570b876-a13b-4b8f-99c4-4900cfe19a21-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-shkn9\" (UID: \"8570b876-a13b-4b8f-99c4-4900cfe19a21\") " pod="openshift-multus/multus-additional-cni-plugins-shkn9" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.532189 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-host-slash\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.532238 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-log-socket\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.532276 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.532298 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/13a2246b-93bb-4586-98ee-53fc84aaae02-os-release\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.532321 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-node-log\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.532336 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/13a2246b-93bb-4586-98ee-53fc84aaae02-multus-conf-dir\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.532351 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zfr5\" (UniqueName: \"kubernetes.io/projected/13a2246b-93bb-4586-98ee-53fc84aaae02-kube-api-access-4zfr5\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.532384 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/13a2246b-93bb-4586-98ee-53fc84aaae02-cni-binary-copy\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.532406 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/13a2246b-93bb-4586-98ee-53fc84aaae02-host-var-lib-cni-bin\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.532427 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/13a2246b-93bb-4586-98ee-53fc84aaae02-hostroot\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.532449 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8570b876-a13b-4b8f-99c4-4900cfe19a21-cni-binary-copy\") pod \"multus-additional-cni-plugins-shkn9\" (UID: \"8570b876-a13b-4b8f-99c4-4900cfe19a21\") " pod="openshift-multus/multus-additional-cni-plugins-shkn9" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.532474 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a308045b-f54d-42ec-8fdf-5bc3e54ef363-env-overrides\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.532493 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/13a2246b-93bb-4586-98ee-53fc84aaae02-system-cni-dir\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.532515 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a308045b-f54d-42ec-8fdf-5bc3e54ef363-ovnkube-config\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.532533 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/13a2246b-93bb-4586-98ee-53fc84aaae02-multus-socket-dir-parent\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.532547 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/13a2246b-93bb-4586-98ee-53fc84aaae02-host-var-lib-cni-multus\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.532567 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-host-cni-bin\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.532616 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8570b876-a13b-4b8f-99c4-4900cfe19a21-os-release\") pod \"multus-additional-cni-plugins-shkn9\" (UID: \"8570b876-a13b-4b8f-99c4-4900cfe19a21\") " pod="openshift-multus/multus-additional-cni-plugins-shkn9" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.532633 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-systemd-units\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.532647 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-run-systemd\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.532666 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/13a2246b-93bb-4586-98ee-53fc84aaae02-etc-kubernetes\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.532681 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-run-openvswitch\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.532816 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-host-run-ovn-kubernetes\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.533040 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/13a2246b-93bb-4586-98ee-53fc84aaae02-multus-cni-dir\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.533055 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/13a2246b-93bb-4586-98ee-53fc84aaae02-host-run-netns\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.533073 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-host-cni-netd\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.533134 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a308045b-f54d-42ec-8fdf-5bc3e54ef363-ovn-node-metrics-cert\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.533178 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/13a2246b-93bb-4586-98ee-53fc84aaae02-cnibin\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.533196 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/13a2246b-93bb-4586-98ee-53fc84aaae02-multus-daemon-config\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.533228 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8570b876-a13b-4b8f-99c4-4900cfe19a21-cnibin\") pod \"multus-additional-cni-plugins-shkn9\" (UID: \"8570b876-a13b-4b8f-99c4-4900cfe19a21\") " pod="openshift-multus/multus-additional-cni-plugins-shkn9" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.533247 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8570b876-a13b-4b8f-99c4-4900cfe19a21-tuning-conf-dir\") pod \"multus-additional-cni-plugins-shkn9\" (UID: \"8570b876-a13b-4b8f-99c4-4900cfe19a21\") " pod="openshift-multus/multus-additional-cni-plugins-shkn9" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.533263 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-host-kubelet\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.533284 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-etc-openvswitch\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.533302 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-var-lib-openvswitch\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.533317 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a308045b-f54d-42ec-8fdf-5bc3e54ef363-ovnkube-script-lib\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.533333 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/13a2246b-93bb-4586-98ee-53fc84aaae02-host-run-k8s-cni-cncf-io\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.535441 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.550982 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef7b114-0e07-402d-a37b-315c36011f4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-scwpp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.568911 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88068a1cf89e2c507b9d93d305db6af5af196bd9eec805f42d91b6fe090aba35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.571245 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-qpr6r" event={"ID":"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0","Type":"ContainerStarted","Data":"3b8f12cd789caf08587b20ffe02edf137dae3e713b7f70d2c36e5af9e65f2dba"} Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.575526 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" event={"ID":"fef7b114-0e07-402d-a37b-315c36011f4b","Type":"ContainerStarted","Data":"b67d090dbd4e672a6d3e777da244480eaf270518c68f8eaf2292ef7c366b8da4"} Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.575586 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" event={"ID":"fef7b114-0e07-402d-a37b-315c36011f4b","Type":"ContainerStarted","Data":"f4b532cef63961a2d5197e711f629f8eabf79582744f74cc715a16a494f7f5f2"} Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.575707 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.587668 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.598930 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.612792 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tpwwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13a2246b-93bb-4586-98ee-53fc84aaae02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zfr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tpwwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.627923 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1c8f149-f188-4759-b2dc-9dc15be1bb13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:40:37.024907 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:40:37.026665 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1609525021/tls.crt::/tmp/serving-cert-1609525021/tls.key\\\\\\\"\\\\nI1201 08:40:42.169070 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:40:42.177896 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:40:42.177940 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:40:42.177968 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:40:42.177975 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:40:42.184767 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1201 08:40:42.184781 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1201 08:40:42.184802 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184808 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184814 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:40:42.184819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:40:42.184822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:40:42.184826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1201 08:40:42.187129 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.634794 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8570b876-a13b-4b8f-99c4-4900cfe19a21-os-release\") pod \"multus-additional-cni-plugins-shkn9\" (UID: \"8570b876-a13b-4b8f-99c4-4900cfe19a21\") " pod="openshift-multus/multus-additional-cni-plugins-shkn9" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.634850 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-systemd-units\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.634873 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-run-systemd\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.634914 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/13a2246b-93bb-4586-98ee-53fc84aaae02-etc-kubernetes\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.634938 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/13a2246b-93bb-4586-98ee-53fc84aaae02-host-run-netns\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.634957 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-run-openvswitch\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.634976 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-host-run-ovn-kubernetes\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.634996 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/13a2246b-93bb-4586-98ee-53fc84aaae02-multus-cni-dir\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635075 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/13a2246b-93bb-4586-98ee-53fc84aaae02-cnibin\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635101 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/13a2246b-93bb-4586-98ee-53fc84aaae02-multus-daemon-config\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635125 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-host-cni-netd\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635146 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a308045b-f54d-42ec-8fdf-5bc3e54ef363-ovn-node-metrics-cert\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635168 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8570b876-a13b-4b8f-99c4-4900cfe19a21-cnibin\") pod \"multus-additional-cni-plugins-shkn9\" (UID: \"8570b876-a13b-4b8f-99c4-4900cfe19a21\") " pod="openshift-multus/multus-additional-cni-plugins-shkn9" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635187 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8570b876-a13b-4b8f-99c4-4900cfe19a21-tuning-conf-dir\") pod \"multus-additional-cni-plugins-shkn9\" (UID: \"8570b876-a13b-4b8f-99c4-4900cfe19a21\") " pod="openshift-multus/multus-additional-cni-plugins-shkn9" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635207 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-host-kubelet\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635230 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-etc-openvswitch\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635253 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/13a2246b-93bb-4586-98ee-53fc84aaae02-host-run-k8s-cni-cncf-io\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635298 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-var-lib-openvswitch\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635323 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a308045b-f54d-42ec-8fdf-5bc3e54ef363-ovnkube-script-lib\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635346 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-run-ovn\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635391 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/13a2246b-93bb-4586-98ee-53fc84aaae02-host-run-multus-certs\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635414 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/13a2246b-93bb-4586-98ee-53fc84aaae02-host-var-lib-kubelet\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635410 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/13a2246b-93bb-4586-98ee-53fc84aaae02-cnibin\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635456 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-host-cni-netd\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635430 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8570b876-a13b-4b8f-99c4-4900cfe19a21-os-release\") pod \"multus-additional-cni-plugins-shkn9\" (UID: \"8570b876-a13b-4b8f-99c4-4900cfe19a21\") " pod="openshift-multus/multus-additional-cni-plugins-shkn9" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635436 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxfd8\" (UniqueName: \"kubernetes.io/projected/a308045b-f54d-42ec-8fdf-5bc3e54ef363-kube-api-access-dxfd8\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635518 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-systemd-units\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635531 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8570b876-a13b-4b8f-99c4-4900cfe19a21-system-cni-dir\") pod \"multus-additional-cni-plugins-shkn9\" (UID: \"8570b876-a13b-4b8f-99c4-4900cfe19a21\") " pod="openshift-multus/multus-additional-cni-plugins-shkn9" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635545 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-run-systemd\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635552 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbkv6\" (UniqueName: \"kubernetes.io/projected/8570b876-a13b-4b8f-99c4-4900cfe19a21-kube-api-access-mbkv6\") pod \"multus-additional-cni-plugins-shkn9\" (UID: \"8570b876-a13b-4b8f-99c4-4900cfe19a21\") " pod="openshift-multus/multus-additional-cni-plugins-shkn9" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635570 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/13a2246b-93bb-4586-98ee-53fc84aaae02-etc-kubernetes\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635575 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-host-run-netns\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635595 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/13a2246b-93bb-4586-98ee-53fc84aaae02-host-run-netns\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635605 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8570b876-a13b-4b8f-99c4-4900cfe19a21-cnibin\") pod \"multus-additional-cni-plugins-shkn9\" (UID: \"8570b876-a13b-4b8f-99c4-4900cfe19a21\") " pod="openshift-multus/multus-additional-cni-plugins-shkn9" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635612 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-host-slash\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635659 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-log-socket\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635679 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635698 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/13a2246b-93bb-4586-98ee-53fc84aaae02-os-release\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635690 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-host-run-ovn-kubernetes\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635738 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/8570b876-a13b-4b8f-99c4-4900cfe19a21-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-shkn9\" (UID: \"8570b876-a13b-4b8f-99c4-4900cfe19a21\") " pod="openshift-multus/multus-additional-cni-plugins-shkn9" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635709 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/13a2246b-93bb-4586-98ee-53fc84aaae02-multus-cni-dir\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635738 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/13a2246b-93bb-4586-98ee-53fc84aaae02-host-run-k8s-cni-cncf-io\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635747 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-run-openvswitch\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635690 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-var-lib-openvswitch\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635776 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-host-slash\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635778 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-host-run-netns\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635812 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-etc-openvswitch\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635819 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-host-kubelet\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635813 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/13a2246b-93bb-4586-98ee-53fc84aaae02-host-var-lib-kubelet\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635832 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635816 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-run-ovn\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635839 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8570b876-a13b-4b8f-99c4-4900cfe19a21-system-cni-dir\") pod \"multus-additional-cni-plugins-shkn9\" (UID: \"8570b876-a13b-4b8f-99c4-4900cfe19a21\") " pod="openshift-multus/multus-additional-cni-plugins-shkn9" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635851 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zfr5\" (UniqueName: \"kubernetes.io/projected/13a2246b-93bb-4586-98ee-53fc84aaae02-kube-api-access-4zfr5\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635868 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/13a2246b-93bb-4586-98ee-53fc84aaae02-host-run-multus-certs\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635862 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-log-socket\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635887 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/13a2246b-93bb-4586-98ee-53fc84aaae02-os-release\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635936 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-node-log\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635965 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/13a2246b-93bb-4586-98ee-53fc84aaae02-multus-conf-dir\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635969 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-node-log\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635990 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/13a2246b-93bb-4586-98ee-53fc84aaae02-multus-conf-dir\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.635997 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8570b876-a13b-4b8f-99c4-4900cfe19a21-tuning-conf-dir\") pod \"multus-additional-cni-plugins-shkn9\" (UID: \"8570b876-a13b-4b8f-99c4-4900cfe19a21\") " pod="openshift-multus/multus-additional-cni-plugins-shkn9" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.636009 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/13a2246b-93bb-4586-98ee-53fc84aaae02-cni-binary-copy\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.636050 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/13a2246b-93bb-4586-98ee-53fc84aaae02-host-var-lib-cni-bin\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.636062 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/13a2246b-93bb-4586-98ee-53fc84aaae02-host-var-lib-cni-bin\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.636138 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/13a2246b-93bb-4586-98ee-53fc84aaae02-hostroot\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.636157 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8570b876-a13b-4b8f-99c4-4900cfe19a21-cni-binary-copy\") pod \"multus-additional-cni-plugins-shkn9\" (UID: \"8570b876-a13b-4b8f-99c4-4900cfe19a21\") " pod="openshift-multus/multus-additional-cni-plugins-shkn9" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.636165 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/13a2246b-93bb-4586-98ee-53fc84aaae02-hostroot\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.636177 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a308045b-f54d-42ec-8fdf-5bc3e54ef363-env-overrides\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.636196 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/13a2246b-93bb-4586-98ee-53fc84aaae02-system-cni-dir\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.636211 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-host-cni-bin\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.636226 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a308045b-f54d-42ec-8fdf-5bc3e54ef363-ovnkube-config\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.636242 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/13a2246b-93bb-4586-98ee-53fc84aaae02-multus-socket-dir-parent\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.636259 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/13a2246b-93bb-4586-98ee-53fc84aaae02-host-var-lib-cni-multus\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.636309 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/13a2246b-93bb-4586-98ee-53fc84aaae02-host-var-lib-cni-multus\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.636323 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/13a2246b-93bb-4586-98ee-53fc84aaae02-system-cni-dir\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.636349 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/13a2246b-93bb-4586-98ee-53fc84aaae02-multus-socket-dir-parent\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.636450 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/13a2246b-93bb-4586-98ee-53fc84aaae02-multus-daemon-config\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.636461 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-host-cni-bin\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.636858 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a308045b-f54d-42ec-8fdf-5bc3e54ef363-ovnkube-script-lib\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.636880 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/13a2246b-93bb-4586-98ee-53fc84aaae02-cni-binary-copy\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.637068 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a308045b-f54d-42ec-8fdf-5bc3e54ef363-ovnkube-config\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.637241 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8570b876-a13b-4b8f-99c4-4900cfe19a21-cni-binary-copy\") pod \"multus-additional-cni-plugins-shkn9\" (UID: \"8570b876-a13b-4b8f-99c4-4900cfe19a21\") " pod="openshift-multus/multus-additional-cni-plugins-shkn9" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.637246 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/8570b876-a13b-4b8f-99c4-4900cfe19a21-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-shkn9\" (UID: \"8570b876-a13b-4b8f-99c4-4900cfe19a21\") " pod="openshift-multus/multus-additional-cni-plugins-shkn9" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.637295 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a308045b-f54d-42ec-8fdf-5bc3e54ef363-env-overrides\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.641468 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a308045b-f54d-42ec-8fdf-5bc3e54ef363-ovn-node-metrics-cert\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.664856 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90f08b6fa658ce25fbc25d8d44d956cdb6424a0b047205f19d1a6fd6870b1e90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8d27e3678f1f63f55bd1b76668e328f49368d67d4d7338a6326cb73d13e08f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.673812 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbkv6\" (UniqueName: \"kubernetes.io/projected/8570b876-a13b-4b8f-99c4-4900cfe19a21-kube-api-access-mbkv6\") pod \"multus-additional-cni-plugins-shkn9\" (UID: \"8570b876-a13b-4b8f-99c4-4900cfe19a21\") " pod="openshift-multus/multus-additional-cni-plugins-shkn9" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.675716 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zfr5\" (UniqueName: \"kubernetes.io/projected/13a2246b-93bb-4586-98ee-53fc84aaae02-kube-api-access-4zfr5\") pod \"multus-tpwwr\" (UID: \"13a2246b-93bb-4586-98ee-53fc84aaae02\") " pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.680405 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxfd8\" (UniqueName: \"kubernetes.io/projected/a308045b-f54d-42ec-8fdf-5bc3e54ef363-kube-api-access-dxfd8\") pod \"ovnkube-node-dxlvk\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.704086 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qpr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qpr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.712253 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-shkn9" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.727433 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-tpwwr" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.729922 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8570b876-a13b-4b8f-99c4-4900cfe19a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shkn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: W1201 08:40:44.738570 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod13a2246b_93bb_4586_98ee_53fc84aaae02.slice/crio-a42ed98026aebd6f364eb1a06bb698d29d535c24165dfe3669897ba0b157a13e WatchSource:0}: Error finding container a42ed98026aebd6f364eb1a06bb698d29d535c24165dfe3669897ba0b157a13e: Status 404 returned error can't find the container with id a42ed98026aebd6f364eb1a06bb698d29d535c24165dfe3669897ba0b157a13e Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.746815 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef7b114-0e07-402d-a37b-315c36011f4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-scwpp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.763004 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f7ac0b4-d048-46e4-bcee-4a271503e46d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8208cd5fda4df2e3b5a8bd90a18c8aca62b16cc2c60e06f52411f24f3d60dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba3cfd8a611c687414d7ecd6b02b423b351d7ac2a4d041085d4ca3bb2b0709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ae6257c73f3632c499b5f107178f64a9cac2f4325c57b41aa75d6d4ea14dd12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5088b5bf7bc847f59ca9efbf400855a70b99f1bbbf207eb93576c3b3eabae5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.784850 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.800225 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.816083 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.833929 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a308045b-f54d-42ec-8fdf-5bc3e54ef363\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dxlvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.846274 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88068a1cf89e2c507b9d93d305db6af5af196bd9eec805f42d91b6fe090aba35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.864438 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.880943 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1c8f149-f188-4759-b2dc-9dc15be1bb13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:40:37.024907 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:40:37.026665 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1609525021/tls.crt::/tmp/serving-cert-1609525021/tls.key\\\\\\\"\\\\nI1201 08:40:42.169070 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:40:42.177896 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:40:42.177940 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:40:42.177968 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:40:42.177975 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:40:42.184767 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1201 08:40:42.184781 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1201 08:40:42.184802 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184808 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184814 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:40:42.184819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:40:42.184822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:40:42.184826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1201 08:40:42.187129 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.894937 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tpwwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13a2246b-93bb-4586-98ee-53fc84aaae02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zfr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tpwwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.912746 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qpr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qpr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.930449 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8570b876-a13b-4b8f-99c4-4900cfe19a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shkn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.944046 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90f08b6fa658ce25fbc25d8d44d956cdb6424a0b047205f19d1a6fd6870b1e90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8d27e3678f1f63f55bd1b76668e328f49368d67d4d7338a6326cb73d13e08f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.957834 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.971046 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4873]: I1201 08:40:44.984922 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef7b114-0e07-402d-a37b-315c36011f4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-scwpp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.000677 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f7ac0b4-d048-46e4-bcee-4a271503e46d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8208cd5fda4df2e3b5a8bd90a18c8aca62b16cc2c60e06f52411f24f3d60dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba3cfd8a611c687414d7ecd6b02b423b351d7ac2a4d041085d4ca3bb2b0709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ae6257c73f3632c499b5f107178f64a9cac2f4325c57b41aa75d6d4ea14dd12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5088b5bf7bc847f59ca9efbf400855a70b99f1bbbf207eb93576c3b3eabae5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.018457 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88068a1cf89e2c507b9d93d305db6af5af196bd9eec805f42d91b6fe090aba35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:45Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.037957 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:45Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.056772 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:45Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.089065 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a308045b-f54d-42ec-8fdf-5bc3e54ef363\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dxlvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:45Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.137387 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.292864 4873 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.295202 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.295241 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.295252 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.295357 4873 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.302242 4873 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.302692 4873 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.303896 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.303926 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.303939 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.303955 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.303968 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:45Z","lastTransitionTime":"2025-12-01T08:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:45 crc kubenswrapper[4873]: E1201 08:40:45.323711 4873 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963d6124-09f6-48b8-b38c-854877aa92e3\\\",\\\"systemUUID\\\":\\\"27c66b0b-2450-400c-b6ea-cdb9cbbf95ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:45Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.327578 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.327632 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.327646 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.327679 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.327694 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:45Z","lastTransitionTime":"2025-12-01T08:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.339003 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 01 08:40:45 crc kubenswrapper[4873]: E1201 08:40:45.339916 4873 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963d6124-09f6-48b8-b38c-854877aa92e3\\\",\\\"systemUUID\\\":\\\"27c66b0b-2450-400c-b6ea-cdb9cbbf95ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:45Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.344554 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.344595 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.344613 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.344630 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.344642 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:45Z","lastTransitionTime":"2025-12-01T08:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:45 crc kubenswrapper[4873]: E1201 08:40:45.357428 4873 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963d6124-09f6-48b8-b38c-854877aa92e3\\\",\\\"systemUUID\\\":\\\"27c66b0b-2450-400c-b6ea-cdb9cbbf95ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:45Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.361233 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.361269 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.361280 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.361296 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.361306 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:45Z","lastTransitionTime":"2025-12-01T08:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.370555 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 01 08:40:45 crc kubenswrapper[4873]: E1201 08:40:45.377435 4873 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963d6124-09f6-48b8-b38c-854877aa92e3\\\",\\\"systemUUID\\\":\\\"27c66b0b-2450-400c-b6ea-cdb9cbbf95ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:45Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.384371 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.384412 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.384425 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.384440 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.384451 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:45Z","lastTransitionTime":"2025-12-01T08:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:45 crc kubenswrapper[4873]: E1201 08:40:45.403064 4873 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963d6124-09f6-48b8-b38c-854877aa92e3\\\",\\\"systemUUID\\\":\\\"27c66b0b-2450-400c-b6ea-cdb9cbbf95ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:45Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:45 crc kubenswrapper[4873]: E1201 08:40:45.403230 4873 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.405240 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.405282 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.405292 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.405306 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.405316 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:45Z","lastTransitionTime":"2025-12-01T08:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.429701 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.432619 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:45 crc kubenswrapper[4873]: W1201 08:40:45.448092 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda308045b_f54d_42ec_8fdf_5bc3e54ef363.slice/crio-c0897dd565e0e5ec410cf9ace85340d1c2f155a4c498b0a8d0d25ace45fa5f65 WatchSource:0}: Error finding container c0897dd565e0e5ec410cf9ace85340d1c2f155a4c498b0a8d0d25ace45fa5f65: Status 404 returned error can't find the container with id c0897dd565e0e5ec410cf9ace85340d1c2f155a4c498b0a8d0d25ace45fa5f65 Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.484749 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.508730 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.508787 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.508800 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.508821 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.508837 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:45Z","lastTransitionTime":"2025-12-01T08:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.579963 4873 generic.go:334] "Generic (PLEG): container finished" podID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerID="a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70" exitCode=0 Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.580069 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" event={"ID":"a308045b-f54d-42ec-8fdf-5bc3e54ef363","Type":"ContainerDied","Data":"a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70"} Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.580123 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" event={"ID":"a308045b-f54d-42ec-8fdf-5bc3e54ef363","Type":"ContainerStarted","Data":"c0897dd565e0e5ec410cf9ace85340d1c2f155a4c498b0a8d0d25ace45fa5f65"} Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.584986 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-qpr6r" event={"ID":"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0","Type":"ContainerStarted","Data":"c42ab88e96b59d9430a69343e06a329447ae0163cb0f5b9ef2bcafa1bf356583"} Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.587791 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-tpwwr" event={"ID":"13a2246b-93bb-4586-98ee-53fc84aaae02","Type":"ContainerStarted","Data":"3dfcc763fe6ab69cab09f09c92ad9a85a4ded49c1443904134f78d1ca0092dae"} Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.587853 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-tpwwr" event={"ID":"13a2246b-93bb-4586-98ee-53fc84aaae02","Type":"ContainerStarted","Data":"a42ed98026aebd6f364eb1a06bb698d29d535c24165dfe3669897ba0b157a13e"} Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.592770 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" event={"ID":"8570b876-a13b-4b8f-99c4-4900cfe19a21","Type":"ContainerStarted","Data":"7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859"} Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.592810 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" event={"ID":"8570b876-a13b-4b8f-99c4-4900cfe19a21","Type":"ContainerStarted","Data":"cf411d56fd6450fa4e1ef26b46bb4b5431db446f74db3c0f3940551bc5ab1ab9"} Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.598384 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f7ac0b4-d048-46e4-bcee-4a271503e46d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8208cd5fda4df2e3b5a8bd90a18c8aca62b16cc2c60e06f52411f24f3d60dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba3cfd8a611c687414d7ecd6b02b423b351d7ac2a4d041085d4ca3bb2b0709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ae6257c73f3632c499b5f107178f64a9cac2f4325c57b41aa75d6d4ea14dd12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5088b5bf7bc847f59ca9efbf400855a70b99f1bbbf207eb93576c3b3eabae5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:45Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.598638 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" event={"ID":"fef7b114-0e07-402d-a37b-315c36011f4b","Type":"ContainerStarted","Data":"26e4a6e0a8877a213e2e974d4f7a3ad84938ae1365fa45ca7621bc050470e7e0"} Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.611864 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.612458 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.612549 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.612618 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.612677 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:45Z","lastTransitionTime":"2025-12-01T08:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.620483 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:45Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.633081 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:45Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.646694 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef7b114-0e07-402d-a37b-315c36011f4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-scwpp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:45Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.648382 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.657534 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.668493 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a308045b-f54d-42ec-8fdf-5bc3e54ef363\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dxlvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:45Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.682731 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88068a1cf89e2c507b9d93d305db6af5af196bd9eec805f42d91b6fe090aba35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:45Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.700708 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:45Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.715943 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.715993 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.716006 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.716045 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.716058 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:45Z","lastTransitionTime":"2025-12-01T08:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.717899 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:45Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.735787 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1c8f149-f188-4759-b2dc-9dc15be1bb13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:40:37.024907 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:40:37.026665 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1609525021/tls.crt::/tmp/serving-cert-1609525021/tls.key\\\\\\\"\\\\nI1201 08:40:42.169070 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:40:42.177896 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:40:42.177940 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:40:42.177968 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:40:42.177975 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:40:42.184767 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1201 08:40:42.184781 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1201 08:40:42.184802 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184808 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184814 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:40:42.184819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:40:42.184822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:40:42.184826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1201 08:40:42.187129 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:45Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.736808 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.751940 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.757403 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.759047 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tpwwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13a2246b-93bb-4586-98ee-53fc84aaae02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zfr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tpwwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:45Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.767917 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.785478 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90f08b6fa658ce25fbc25d8d44d956cdb6424a0b047205f19d1a6fd6870b1e90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8d27e3678f1f63f55bd1b76668e328f49368d67d4d7338a6326cb73d13e08f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:45Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.787539 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.798855 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qpr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qpr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:45Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.815231 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8570b876-a13b-4b8f-99c4-4900cfe19a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shkn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:45Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.820348 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.820401 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.820414 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.820436 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.820451 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:45Z","lastTransitionTime":"2025-12-01T08:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.829532 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90f08b6fa658ce25fbc25d8d44d956cdb6424a0b047205f19d1a6fd6870b1e90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8d27e3678f1f63f55bd1b76668e328f49368d67d4d7338a6326cb73d13e08f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:45Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.851543 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qpr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c42ab88e96b59d9430a69343e06a329447ae0163cb0f5b9ef2bcafa1bf356583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qpr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:45Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.888332 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8570b876-a13b-4b8f-99c4-4900cfe19a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shkn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:45Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.923764 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.923818 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.923831 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.923852 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.923865 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:45Z","lastTransitionTime":"2025-12-01T08:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.926638 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f7ac0b4-d048-46e4-bcee-4a271503e46d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8208cd5fda4df2e3b5a8bd90a18c8aca62b16cc2c60e06f52411f24f3d60dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba3cfd8a611c687414d7ecd6b02b423b351d7ac2a4d041085d4ca3bb2b0709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ae6257c73f3632c499b5f107178f64a9cac2f4325c57b41aa75d6d4ea14dd12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5088b5bf7bc847f59ca9efbf400855a70b99f1bbbf207eb93576c3b3eabae5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:45Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:45 crc kubenswrapper[4873]: I1201 08:40:45.966348 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:45Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.005358 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.026077 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.026132 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.026147 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.026170 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.026185 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:46Z","lastTransitionTime":"2025-12-01T08:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.044394 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef7b114-0e07-402d-a37b-315c36011f4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26e4a6e0a8877a213e2e974d4f7a3ad84938ae1365fa45ca7621bc050470e7e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67d090dbd4e672a6d3e777da244480eaf270518c68f8eaf2292ef7c366b8da4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-scwpp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.090276 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a308045b-f54d-42ec-8fdf-5bc3e54ef363\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dxlvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.128516 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.128581 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.128591 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.128626 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.128641 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:46Z","lastTransitionTime":"2025-12-01T08:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.135948 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a2348f-009e-4cc1-b591-e1c14ff0faf2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff98ccdc9925e44e62f0d6e826ea7926dfa2a7d83d33e475a092f128b7fe920a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4676d506268107b87ebe2e21ba6e65d959a23a4d6f38bf0de588bd9c5d15e207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4751109230b67ad8948a2de4794571581d662bfd8abc6dc5c198dbc411172860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3acd19dc33ad40d8857d28bc0ccee4c99a6927e92e5029873234ecb5dc6ed18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7242d1034bd1938cb0bd750cab74db52d454ddd8af75f1e9e0a3c007dcdddd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.155386 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.155482 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.155508 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:40:46 crc kubenswrapper[4873]: E1201 08:40:46.155565 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:40:50.155540184 +0000 UTC m=+26.057648713 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:40:46 crc kubenswrapper[4873]: E1201 08:40:46.155586 4873 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:40:46 crc kubenswrapper[4873]: E1201 08:40:46.155661 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:40:50.155641007 +0000 UTC m=+26.057749536 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.155680 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.155705 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:40:46 crc kubenswrapper[4873]: E1201 08:40:46.155767 4873 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:40:46 crc kubenswrapper[4873]: E1201 08:40:46.155815 4873 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:40:46 crc kubenswrapper[4873]: E1201 08:40:46.155832 4873 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:40:46 crc kubenswrapper[4873]: E1201 08:40:46.155844 4873 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:46 crc kubenswrapper[4873]: E1201 08:40:46.155867 4873 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:40:46 crc kubenswrapper[4873]: E1201 08:40:46.155890 4873 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:40:46 crc kubenswrapper[4873]: E1201 08:40:46.155902 4873 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:46 crc kubenswrapper[4873]: E1201 08:40:46.155868 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 08:40:50.155862022 +0000 UTC m=+26.057970561 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:46 crc kubenswrapper[4873]: E1201 08:40:46.155966 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:40:50.155953025 +0000 UTC m=+26.058061564 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:40:46 crc kubenswrapper[4873]: E1201 08:40:46.155981 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 08:40:50.155974515 +0000 UTC m=+26.058083054 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.168898 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88068a1cf89e2c507b9d93d305db6af5af196bd9eec805f42d91b6fe090aba35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.208283 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.230968 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.231053 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.231073 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.231114 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.231135 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:46Z","lastTransitionTime":"2025-12-01T08:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.248962 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.291164 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1c8f149-f188-4759-b2dc-9dc15be1bb13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:40:37.024907 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:40:37.026665 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1609525021/tls.crt::/tmp/serving-cert-1609525021/tls.key\\\\\\\"\\\\nI1201 08:40:42.169070 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:40:42.177896 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:40:42.177940 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:40:42.177968 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:40:42.177975 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:40:42.184767 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1201 08:40:42.184781 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1201 08:40:42.184802 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184808 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184814 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:40:42.184819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:40:42.184822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:40:42.184826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1201 08:40:42.187129 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.327187 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tpwwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13a2246b-93bb-4586-98ee-53fc84aaae02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dfcc763fe6ab69cab09f09c92ad9a85a4ded49c1443904134f78d1ca0092dae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zfr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tpwwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.334076 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.334119 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.334132 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.334152 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.334166 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:46Z","lastTransitionTime":"2025-12-01T08:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.429624 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:40:46 crc kubenswrapper[4873]: E1201 08:40:46.429839 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.430252 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:40:46 crc kubenswrapper[4873]: E1201 08:40:46.430326 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.430367 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:40:46 crc kubenswrapper[4873]: E1201 08:40:46.430442 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.436254 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.436325 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.436344 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.436387 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.436409 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:46Z","lastTransitionTime":"2025-12-01T08:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.539298 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.539346 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.539360 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.539377 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.539388 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:46Z","lastTransitionTime":"2025-12-01T08:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.607453 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" event={"ID":"a308045b-f54d-42ec-8fdf-5bc3e54ef363","Type":"ContainerStarted","Data":"d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14"} Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.607558 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" event={"ID":"a308045b-f54d-42ec-8fdf-5bc3e54ef363","Type":"ContainerStarted","Data":"000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b"} Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.607592 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" event={"ID":"a308045b-f54d-42ec-8fdf-5bc3e54ef363","Type":"ContainerStarted","Data":"61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a"} Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.607623 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" event={"ID":"a308045b-f54d-42ec-8fdf-5bc3e54ef363","Type":"ContainerStarted","Data":"26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c"} Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.607666 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" event={"ID":"a308045b-f54d-42ec-8fdf-5bc3e54ef363","Type":"ContainerStarted","Data":"b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d"} Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.607688 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" event={"ID":"a308045b-f54d-42ec-8fdf-5bc3e54ef363","Type":"ContainerStarted","Data":"d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff"} Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.609132 4873 generic.go:334] "Generic (PLEG): container finished" podID="8570b876-a13b-4b8f-99c4-4900cfe19a21" containerID="7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859" exitCode=0 Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.609219 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" event={"ID":"8570b876-a13b-4b8f-99c4-4900cfe19a21","Type":"ContainerDied","Data":"7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859"} Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.611503 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"b36539974b64191f4ca087bacbd02333f90c89fc8273ab7285e9a1d840155707"} Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.635808 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1c8f149-f188-4759-b2dc-9dc15be1bb13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:40:37.024907 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:40:37.026665 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1609525021/tls.crt::/tmp/serving-cert-1609525021/tls.key\\\\\\\"\\\\nI1201 08:40:42.169070 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:40:42.177896 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:40:42.177940 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:40:42.177968 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:40:42.177975 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:40:42.184767 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1201 08:40:42.184781 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1201 08:40:42.184802 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184808 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184814 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:40:42.184819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:40:42.184822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:40:42.184826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1201 08:40:42.187129 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.642673 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.642719 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.642734 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.642754 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.642767 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:46Z","lastTransitionTime":"2025-12-01T08:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.657608 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tpwwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13a2246b-93bb-4586-98ee-53fc84aaae02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dfcc763fe6ab69cab09f09c92ad9a85a4ded49c1443904134f78d1ca0092dae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zfr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tpwwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.682760 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90f08b6fa658ce25fbc25d8d44d956cdb6424a0b047205f19d1a6fd6870b1e90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8d27e3678f1f63f55bd1b76668e328f49368d67d4d7338a6326cb73d13e08f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.695391 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qpr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c42ab88e96b59d9430a69343e06a329447ae0163cb0f5b9ef2bcafa1bf356583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qpr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.714875 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8570b876-a13b-4b8f-99c4-4900cfe19a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shkn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.729721 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f7ac0b4-d048-46e4-bcee-4a271503e46d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8208cd5fda4df2e3b5a8bd90a18c8aca62b16cc2c60e06f52411f24f3d60dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba3cfd8a611c687414d7ecd6b02b423b351d7ac2a4d041085d4ca3bb2b0709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ae6257c73f3632c499b5f107178f64a9cac2f4325c57b41aa75d6d4ea14dd12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5088b5bf7bc847f59ca9efbf400855a70b99f1bbbf207eb93576c3b3eabae5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.744952 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.746468 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.746516 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.746530 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.746548 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.746567 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:46Z","lastTransitionTime":"2025-12-01T08:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.759219 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.774390 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef7b114-0e07-402d-a37b-315c36011f4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26e4a6e0a8877a213e2e974d4f7a3ad84938ae1365fa45ca7621bc050470e7e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67d090dbd4e672a6d3e777da244480eaf270518c68f8eaf2292ef7c366b8da4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-scwpp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.797760 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a2348f-009e-4cc1-b591-e1c14ff0faf2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff98ccdc9925e44e62f0d6e826ea7926dfa2a7d83d33e475a092f128b7fe920a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4676d506268107b87ebe2e21ba6e65d959a23a4d6f38bf0de588bd9c5d15e207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4751109230b67ad8948a2de4794571581d662bfd8abc6dc5c198dbc411172860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3acd19dc33ad40d8857d28bc0ccee4c99a6927e92e5029873234ecb5dc6ed18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7242d1034bd1938cb0bd750cab74db52d454ddd8af75f1e9e0a3c007dcdddd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.809959 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88068a1cf89e2c507b9d93d305db6af5af196bd9eec805f42d91b6fe090aba35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.826552 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.850101 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.850147 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.850156 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.850174 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.850185 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:46Z","lastTransitionTime":"2025-12-01T08:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.868836 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.895140 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a308045b-f54d-42ec-8fdf-5bc3e54ef363\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dxlvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.932357 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a2348f-009e-4cc1-b591-e1c14ff0faf2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff98ccdc9925e44e62f0d6e826ea7926dfa2a7d83d33e475a092f128b7fe920a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4676d506268107b87ebe2e21ba6e65d959a23a4d6f38bf0de588bd9c5d15e207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4751109230b67ad8948a2de4794571581d662bfd8abc6dc5c198dbc411172860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3acd19dc33ad40d8857d28bc0ccee4c99a6927e92e5029873234ecb5dc6ed18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7242d1034bd1938cb0bd750cab74db52d454ddd8af75f1e9e0a3c007dcdddd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.952966 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.953044 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.953054 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.953070 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.953079 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:46Z","lastTransitionTime":"2025-12-01T08:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:46 crc kubenswrapper[4873]: I1201 08:40:46.967290 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88068a1cf89e2c507b9d93d305db6af5af196bd9eec805f42d91b6fe090aba35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.007585 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.047345 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.055542 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.055610 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.055628 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.055713 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.055741 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:47Z","lastTransitionTime":"2025-12-01T08:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.095197 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a308045b-f54d-42ec-8fdf-5bc3e54ef363\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dxlvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.126087 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1c8f149-f188-4759-b2dc-9dc15be1bb13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:40:37.024907 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:40:37.026665 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1609525021/tls.crt::/tmp/serving-cert-1609525021/tls.key\\\\\\\"\\\\nI1201 08:40:42.169070 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:40:42.177896 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:40:42.177940 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:40:42.177968 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:40:42.177975 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:40:42.184767 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1201 08:40:42.184781 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1201 08:40:42.184802 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184808 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184814 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:40:42.184819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:40:42.184822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:40:42.184826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1201 08:40:42.187129 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.158071 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.158113 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.158126 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.158144 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.158155 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:47Z","lastTransitionTime":"2025-12-01T08:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.170580 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tpwwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13a2246b-93bb-4586-98ee-53fc84aaae02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dfcc763fe6ab69cab09f09c92ad9a85a4ded49c1443904134f78d1ca0092dae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zfr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tpwwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.211448 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90f08b6fa658ce25fbc25d8d44d956cdb6424a0b047205f19d1a6fd6870b1e90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8d27e3678f1f63f55bd1b76668e328f49368d67d4d7338a6326cb73d13e08f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.247633 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qpr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c42ab88e96b59d9430a69343e06a329447ae0163cb0f5b9ef2bcafa1bf356583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qpr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.260618 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.260660 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.260670 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.260686 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.260697 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:47Z","lastTransitionTime":"2025-12-01T08:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.291697 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8570b876-a13b-4b8f-99c4-4900cfe19a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shkn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.333114 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f7ac0b4-d048-46e4-bcee-4a271503e46d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8208cd5fda4df2e3b5a8bd90a18c8aca62b16cc2c60e06f52411f24f3d60dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba3cfd8a611c687414d7ecd6b02b423b351d7ac2a4d041085d4ca3bb2b0709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ae6257c73f3632c499b5f107178f64a9cac2f4325c57b41aa75d6d4ea14dd12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5088b5bf7bc847f59ca9efbf400855a70b99f1bbbf207eb93576c3b3eabae5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.363976 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.364034 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.364044 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.364065 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.364079 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:47Z","lastTransitionTime":"2025-12-01T08:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.368960 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.412877 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36539974b64191f4ca087bacbd02333f90c89fc8273ab7285e9a1d840155707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.451282 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef7b114-0e07-402d-a37b-315c36011f4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26e4a6e0a8877a213e2e974d4f7a3ad84938ae1365fa45ca7621bc050470e7e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67d090dbd4e672a6d3e777da244480eaf270518c68f8eaf2292ef7c366b8da4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-scwpp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.467178 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.467250 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.467270 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.467297 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.467317 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:47Z","lastTransitionTime":"2025-12-01T08:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.570664 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.570723 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.570740 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.570765 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.570782 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:47Z","lastTransitionTime":"2025-12-01T08:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.618208 4873 generic.go:334] "Generic (PLEG): container finished" podID="8570b876-a13b-4b8f-99c4-4900cfe19a21" containerID="fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023" exitCode=0 Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.618269 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" event={"ID":"8570b876-a13b-4b8f-99c4-4900cfe19a21","Type":"ContainerDied","Data":"fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023"} Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.636351 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qpr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c42ab88e96b59d9430a69343e06a329447ae0163cb0f5b9ef2bcafa1bf356583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qpr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.674408 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.674465 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.674480 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.674503 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.674521 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:47Z","lastTransitionTime":"2025-12-01T08:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.676520 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8570b876-a13b-4b8f-99c4-4900cfe19a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shkn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.696950 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90f08b6fa658ce25fbc25d8d44d956cdb6424a0b047205f19d1a6fd6870b1e90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8d27e3678f1f63f55bd1b76668e328f49368d67d4d7338a6326cb73d13e08f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.714422 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.729256 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36539974b64191f4ca087bacbd02333f90c89fc8273ab7285e9a1d840155707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.741329 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef7b114-0e07-402d-a37b-315c36011f4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26e4a6e0a8877a213e2e974d4f7a3ad84938ae1365fa45ca7621bc050470e7e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67d090dbd4e672a6d3e777da244480eaf270518c68f8eaf2292ef7c366b8da4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-scwpp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.760061 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f7ac0b4-d048-46e4-bcee-4a271503e46d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8208cd5fda4df2e3b5a8bd90a18c8aca62b16cc2c60e06f52411f24f3d60dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba3cfd8a611c687414d7ecd6b02b423b351d7ac2a4d041085d4ca3bb2b0709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ae6257c73f3632c499b5f107178f64a9cac2f4325c57b41aa75d6d4ea14dd12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5088b5bf7bc847f59ca9efbf400855a70b99f1bbbf207eb93576c3b3eabae5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.760938 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-pfqcc"] Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.761538 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-pfqcc" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.764240 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.777940 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.778965 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.779072 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.779084 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.779098 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.779162 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:47Z","lastTransitionTime":"2025-12-01T08:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.797125 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.817629 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.848778 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88068a1cf89e2c507b9d93d305db6af5af196bd9eec805f42d91b6fe090aba35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.873384 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/06182c66-4be6-4e00-9a17-4917badfc997-host\") pod \"node-ca-pfqcc\" (UID: \"06182c66-4be6-4e00-9a17-4917badfc997\") " pod="openshift-image-registry/node-ca-pfqcc" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.873440 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/06182c66-4be6-4e00-9a17-4917badfc997-serviceca\") pod \"node-ca-pfqcc\" (UID: \"06182c66-4be6-4e00-9a17-4917badfc997\") " pod="openshift-image-registry/node-ca-pfqcc" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.873492 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kv2dp\" (UniqueName: \"kubernetes.io/projected/06182c66-4be6-4e00-9a17-4917badfc997-kube-api-access-kv2dp\") pod \"node-ca-pfqcc\" (UID: \"06182c66-4be6-4e00-9a17-4917badfc997\") " pod="openshift-image-registry/node-ca-pfqcc" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.881947 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.881988 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.882029 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.882050 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.882062 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:47Z","lastTransitionTime":"2025-12-01T08:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.887488 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.927075 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.971665 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a308045b-f54d-42ec-8fdf-5bc3e54ef363\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dxlvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.975063 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/06182c66-4be6-4e00-9a17-4917badfc997-host\") pod \"node-ca-pfqcc\" (UID: \"06182c66-4be6-4e00-9a17-4917badfc997\") " pod="openshift-image-registry/node-ca-pfqcc" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.975128 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/06182c66-4be6-4e00-9a17-4917badfc997-serviceca\") pod \"node-ca-pfqcc\" (UID: \"06182c66-4be6-4e00-9a17-4917badfc997\") " pod="openshift-image-registry/node-ca-pfqcc" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.975184 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kv2dp\" (UniqueName: \"kubernetes.io/projected/06182c66-4be6-4e00-9a17-4917badfc997-kube-api-access-kv2dp\") pod \"node-ca-pfqcc\" (UID: \"06182c66-4be6-4e00-9a17-4917badfc997\") " pod="openshift-image-registry/node-ca-pfqcc" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.975234 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/06182c66-4be6-4e00-9a17-4917badfc997-host\") pod \"node-ca-pfqcc\" (UID: \"06182c66-4be6-4e00-9a17-4917badfc997\") " pod="openshift-image-registry/node-ca-pfqcc" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.976691 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/06182c66-4be6-4e00-9a17-4917badfc997-serviceca\") pod \"node-ca-pfqcc\" (UID: \"06182c66-4be6-4e00-9a17-4917badfc997\") " pod="openshift-image-registry/node-ca-pfqcc" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.984600 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.984652 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.984665 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.984683 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:47 crc kubenswrapper[4873]: I1201 08:40:47.984698 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:47Z","lastTransitionTime":"2025-12-01T08:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.018668 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kv2dp\" (UniqueName: \"kubernetes.io/projected/06182c66-4be6-4e00-9a17-4917badfc997-kube-api-access-kv2dp\") pod \"node-ca-pfqcc\" (UID: \"06182c66-4be6-4e00-9a17-4917badfc997\") " pod="openshift-image-registry/node-ca-pfqcc" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.034664 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a2348f-009e-4cc1-b591-e1c14ff0faf2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff98ccdc9925e44e62f0d6e826ea7926dfa2a7d83d33e475a092f128b7fe920a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4676d506268107b87ebe2e21ba6e65d959a23a4d6f38bf0de588bd9c5d15e207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4751109230b67ad8948a2de4794571581d662bfd8abc6dc5c198dbc411172860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3acd19dc33ad40d8857d28bc0ccee4c99a6927e92e5029873234ecb5dc6ed18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7242d1034bd1938cb0bd750cab74db52d454ddd8af75f1e9e0a3c007dcdddd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.067367 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1c8f149-f188-4759-b2dc-9dc15be1bb13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:40:37.024907 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:40:37.026665 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1609525021/tls.crt::/tmp/serving-cert-1609525021/tls.key\\\\\\\"\\\\nI1201 08:40:42.169070 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:40:42.177896 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:40:42.177940 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:40:42.177968 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:40:42.177975 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:40:42.184767 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1201 08:40:42.184781 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1201 08:40:42.184802 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184808 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184814 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:40:42.184819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:40:42.184822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:40:42.184826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1201 08:40:42.187129 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.085261 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-pfqcc" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.087165 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.087201 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.087214 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.087233 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.087247 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:48Z","lastTransitionTime":"2025-12-01T08:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:48 crc kubenswrapper[4873]: W1201 08:40:48.103500 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod06182c66_4be6_4e00_9a17_4917badfc997.slice/crio-2ab1b023ddaa88a03eb90ce2333450080b0e16a0a3c9d6602f0d151dca70f1fc WatchSource:0}: Error finding container 2ab1b023ddaa88a03eb90ce2333450080b0e16a0a3c9d6602f0d151dca70f1fc: Status 404 returned error can't find the container with id 2ab1b023ddaa88a03eb90ce2333450080b0e16a0a3c9d6602f0d151dca70f1fc Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.111044 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tpwwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13a2246b-93bb-4586-98ee-53fc84aaae02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dfcc763fe6ab69cab09f09c92ad9a85a4ded49c1443904134f78d1ca0092dae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zfr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tpwwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.153394 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1c8f149-f188-4759-b2dc-9dc15be1bb13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:40:37.024907 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:40:37.026665 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1609525021/tls.crt::/tmp/serving-cert-1609525021/tls.key\\\\\\\"\\\\nI1201 08:40:42.169070 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:40:42.177896 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:40:42.177940 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:40:42.177968 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:40:42.177975 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:40:42.184767 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1201 08:40:42.184781 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1201 08:40:42.184802 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184808 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184814 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:40:42.184819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:40:42.184822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:40:42.184826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1201 08:40:42.187129 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.190849 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.190894 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.190908 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.190925 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.190939 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:48Z","lastTransitionTime":"2025-12-01T08:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.191541 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tpwwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13a2246b-93bb-4586-98ee-53fc84aaae02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dfcc763fe6ab69cab09f09c92ad9a85a4ded49c1443904134f78d1ca0092dae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zfr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tpwwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.223596 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qpr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c42ab88e96b59d9430a69343e06a329447ae0163cb0f5b9ef2bcafa1bf356583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qpr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.265767 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8570b876-a13b-4b8f-99c4-4900cfe19a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shkn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.294815 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.294859 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.294869 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.294885 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.294895 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:48Z","lastTransitionTime":"2025-12-01T08:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.305192 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90f08b6fa658ce25fbc25d8d44d956cdb6424a0b047205f19d1a6fd6870b1e90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8d27e3678f1f63f55bd1b76668e328f49368d67d4d7338a6326cb73d13e08f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.346341 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.385771 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36539974b64191f4ca087bacbd02333f90c89fc8273ab7285e9a1d840155707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.398791 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.399424 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.399478 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.399512 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.399524 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:48Z","lastTransitionTime":"2025-12-01T08:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.426841 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef7b114-0e07-402d-a37b-315c36011f4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26e4a6e0a8877a213e2e974d4f7a3ad84938ae1365fa45ca7621bc050470e7e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67d090dbd4e672a6d3e777da244480eaf270518c68f8eaf2292ef7c366b8da4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-scwpp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.428973 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.429051 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.429099 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:40:48 crc kubenswrapper[4873]: E1201 08:40:48.429182 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:40:48 crc kubenswrapper[4873]: E1201 08:40:48.429349 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:40:48 crc kubenswrapper[4873]: E1201 08:40:48.429437 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.463991 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfqcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06182c66-4be6-4e00-9a17-4917badfc997\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfqcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.502542 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.502579 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.502590 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.502604 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.502616 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:48Z","lastTransitionTime":"2025-12-01T08:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.506906 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f7ac0b4-d048-46e4-bcee-4a271503e46d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8208cd5fda4df2e3b5a8bd90a18c8aca62b16cc2c60e06f52411f24f3d60dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba3cfd8a611c687414d7ecd6b02b423b351d7ac2a4d041085d4ca3bb2b0709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ae6257c73f3632c499b5f107178f64a9cac2f4325c57b41aa75d6d4ea14dd12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5088b5bf7bc847f59ca9efbf400855a70b99f1bbbf207eb93576c3b3eabae5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.546331 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88068a1cf89e2c507b9d93d305db6af5af196bd9eec805f42d91b6fe090aba35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.590979 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.605462 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.605503 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.605512 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.605527 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.605537 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:48Z","lastTransitionTime":"2025-12-01T08:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.627502 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" event={"ID":"a308045b-f54d-42ec-8fdf-5bc3e54ef363","Type":"ContainerStarted","Data":"3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c"} Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.631006 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-pfqcc" event={"ID":"06182c66-4be6-4e00-9a17-4917badfc997","Type":"ContainerStarted","Data":"3aa8ef5ef07129ec4a7b4cf4738a01cd6444e9bf548134a9da493c7149301387"} Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.631087 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-pfqcc" event={"ID":"06182c66-4be6-4e00-9a17-4917badfc997","Type":"ContainerStarted","Data":"2ab1b023ddaa88a03eb90ce2333450080b0e16a0a3c9d6602f0d151dca70f1fc"} Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.634072 4873 generic.go:334] "Generic (PLEG): container finished" podID="8570b876-a13b-4b8f-99c4-4900cfe19a21" containerID="5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1" exitCode=0 Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.634132 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" event={"ID":"8570b876-a13b-4b8f-99c4-4900cfe19a21","Type":"ContainerDied","Data":"5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1"} Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.659665 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.688660 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a308045b-f54d-42ec-8fdf-5bc3e54ef363\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dxlvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.708100 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.708144 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.708157 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.708175 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.708210 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:48Z","lastTransitionTime":"2025-12-01T08:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.716512 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a2348f-009e-4cc1-b591-e1c14ff0faf2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff98ccdc9925e44e62f0d6e826ea7926dfa2a7d83d33e475a092f128b7fe920a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4676d506268107b87ebe2e21ba6e65d959a23a4d6f38bf0de588bd9c5d15e207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4751109230b67ad8948a2de4794571581d662bfd8abc6dc5c198dbc411172860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3acd19dc33ad40d8857d28bc0ccee4c99a6927e92e5029873234ecb5dc6ed18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7242d1034bd1938cb0bd750cab74db52d454ddd8af75f1e9e0a3c007dcdddd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.755065 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a2348f-009e-4cc1-b591-e1c14ff0faf2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff98ccdc9925e44e62f0d6e826ea7926dfa2a7d83d33e475a092f128b7fe920a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4676d506268107b87ebe2e21ba6e65d959a23a4d6f38bf0de588bd9c5d15e207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4751109230b67ad8948a2de4794571581d662bfd8abc6dc5c198dbc411172860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3acd19dc33ad40d8857d28bc0ccee4c99a6927e92e5029873234ecb5dc6ed18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7242d1034bd1938cb0bd750cab74db52d454ddd8af75f1e9e0a3c007dcdddd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.788071 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88068a1cf89e2c507b9d93d305db6af5af196bd9eec805f42d91b6fe090aba35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.810901 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.810927 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.810936 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.810951 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.810960 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:48Z","lastTransitionTime":"2025-12-01T08:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.827283 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.866494 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.912401 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a308045b-f54d-42ec-8fdf-5bc3e54ef363\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dxlvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.914132 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.914174 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.914185 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.914205 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.914216 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:48Z","lastTransitionTime":"2025-12-01T08:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.947088 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1c8f149-f188-4759-b2dc-9dc15be1bb13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:40:37.024907 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:40:37.026665 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1609525021/tls.crt::/tmp/serving-cert-1609525021/tls.key\\\\\\\"\\\\nI1201 08:40:42.169070 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:40:42.177896 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:40:42.177940 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:40:42.177968 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:40:42.177975 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:40:42.184767 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1201 08:40:42.184781 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1201 08:40:42.184802 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184808 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184814 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:40:42.184819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:40:42.184822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:40:42.184826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1201 08:40:42.187129 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:48 crc kubenswrapper[4873]: I1201 08:40:48.984800 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tpwwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13a2246b-93bb-4586-98ee-53fc84aaae02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dfcc763fe6ab69cab09f09c92ad9a85a4ded49c1443904134f78d1ca0092dae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zfr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tpwwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.016474 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.016532 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.016550 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.016609 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.016629 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:49Z","lastTransitionTime":"2025-12-01T08:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.024490 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90f08b6fa658ce25fbc25d8d44d956cdb6424a0b047205f19d1a6fd6870b1e90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8d27e3678f1f63f55bd1b76668e328f49368d67d4d7338a6326cb73d13e08f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.063606 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qpr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c42ab88e96b59d9430a69343e06a329447ae0163cb0f5b9ef2bcafa1bf356583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qpr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.105642 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8570b876-a13b-4b8f-99c4-4900cfe19a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shkn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.118507 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.118558 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.118567 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.118579 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.118593 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:49Z","lastTransitionTime":"2025-12-01T08:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.184865 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f7ac0b4-d048-46e4-bcee-4a271503e46d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8208cd5fda4df2e3b5a8bd90a18c8aca62b16cc2c60e06f52411f24f3d60dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba3cfd8a611c687414d7ecd6b02b423b351d7ac2a4d041085d4ca3bb2b0709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ae6257c73f3632c499b5f107178f64a9cac2f4325c57b41aa75d6d4ea14dd12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5088b5bf7bc847f59ca9efbf400855a70b99f1bbbf207eb93576c3b3eabae5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.201945 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.222223 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.222272 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.222285 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.222303 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.222319 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:49Z","lastTransitionTime":"2025-12-01T08:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.231640 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36539974b64191f4ca087bacbd02333f90c89fc8273ab7285e9a1d840155707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.274284 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef7b114-0e07-402d-a37b-315c36011f4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26e4a6e0a8877a213e2e974d4f7a3ad84938ae1365fa45ca7621bc050470e7e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67d090dbd4e672a6d3e777da244480eaf270518c68f8eaf2292ef7c366b8da4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-scwpp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.304767 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfqcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06182c66-4be6-4e00-9a17-4917badfc997\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa8ef5ef07129ec4a7b4cf4738a01cd6444e9bf548134a9da493c7149301387\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfqcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.325515 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.325559 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.325606 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.325625 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.325637 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:49Z","lastTransitionTime":"2025-12-01T08:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.428496 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.428537 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.428554 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.428576 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.428593 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:49Z","lastTransitionTime":"2025-12-01T08:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.531183 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.531375 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.531401 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.531432 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.531469 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:49Z","lastTransitionTime":"2025-12-01T08:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.634981 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.635068 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.635088 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.635112 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.635160 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:49Z","lastTransitionTime":"2025-12-01T08:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.640438 4873 generic.go:334] "Generic (PLEG): container finished" podID="8570b876-a13b-4b8f-99c4-4900cfe19a21" containerID="7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e" exitCode=0 Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.640487 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" event={"ID":"8570b876-a13b-4b8f-99c4-4900cfe19a21","Type":"ContainerDied","Data":"7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e"} Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.665110 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f7ac0b4-d048-46e4-bcee-4a271503e46d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8208cd5fda4df2e3b5a8bd90a18c8aca62b16cc2c60e06f52411f24f3d60dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba3cfd8a611c687414d7ecd6b02b423b351d7ac2a4d041085d4ca3bb2b0709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ae6257c73f3632c499b5f107178f64a9cac2f4325c57b41aa75d6d4ea14dd12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5088b5bf7bc847f59ca9efbf400855a70b99f1bbbf207eb93576c3b3eabae5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.677739 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.688049 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36539974b64191f4ca087bacbd02333f90c89fc8273ab7285e9a1d840155707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.696947 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef7b114-0e07-402d-a37b-315c36011f4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26e4a6e0a8877a213e2e974d4f7a3ad84938ae1365fa45ca7621bc050470e7e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67d090dbd4e672a6d3e777da244480eaf270518c68f8eaf2292ef7c366b8da4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-scwpp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.707442 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfqcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06182c66-4be6-4e00-9a17-4917badfc997\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa8ef5ef07129ec4a7b4cf4738a01cd6444e9bf548134a9da493c7149301387\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfqcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.725437 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a2348f-009e-4cc1-b591-e1c14ff0faf2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff98ccdc9925e44e62f0d6e826ea7926dfa2a7d83d33e475a092f128b7fe920a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4676d506268107b87ebe2e21ba6e65d959a23a4d6f38bf0de588bd9c5d15e207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4751109230b67ad8948a2de4794571581d662bfd8abc6dc5c198dbc411172860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3acd19dc33ad40d8857d28bc0ccee4c99a6927e92e5029873234ecb5dc6ed18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7242d1034bd1938cb0bd750cab74db52d454ddd8af75f1e9e0a3c007dcdddd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.737920 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.737965 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.737978 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.737994 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.738007 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:49Z","lastTransitionTime":"2025-12-01T08:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.741179 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88068a1cf89e2c507b9d93d305db6af5af196bd9eec805f42d91b6fe090aba35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.754182 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.768312 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.786566 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a308045b-f54d-42ec-8fdf-5bc3e54ef363\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dxlvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.801438 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1c8f149-f188-4759-b2dc-9dc15be1bb13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:40:37.024907 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:40:37.026665 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1609525021/tls.crt::/tmp/serving-cert-1609525021/tls.key\\\\\\\"\\\\nI1201 08:40:42.169070 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:40:42.177896 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:40:42.177940 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:40:42.177968 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:40:42.177975 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:40:42.184767 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1201 08:40:42.184781 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1201 08:40:42.184802 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184808 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184814 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:40:42.184819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:40:42.184822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:40:42.184826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1201 08:40:42.187129 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.816612 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tpwwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13a2246b-93bb-4586-98ee-53fc84aaae02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dfcc763fe6ab69cab09f09c92ad9a85a4ded49c1443904134f78d1ca0092dae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zfr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tpwwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.829894 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90f08b6fa658ce25fbc25d8d44d956cdb6424a0b047205f19d1a6fd6870b1e90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8d27e3678f1f63f55bd1b76668e328f49368d67d4d7338a6326cb73d13e08f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.842503 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.842540 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.842550 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.842565 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.842575 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:49Z","lastTransitionTime":"2025-12-01T08:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.863207 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qpr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c42ab88e96b59d9430a69343e06a329447ae0163cb0f5b9ef2bcafa1bf356583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qpr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.912981 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8570b876-a13b-4b8f-99c4-4900cfe19a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shkn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.944927 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.944980 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.944990 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.945005 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:49 crc kubenswrapper[4873]: I1201 08:40:49.945039 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:49Z","lastTransitionTime":"2025-12-01T08:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.048041 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.048130 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.048153 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.048176 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.048193 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:50Z","lastTransitionTime":"2025-12-01T08:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.150855 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.150913 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.150924 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.150956 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.150967 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:50Z","lastTransitionTime":"2025-12-01T08:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.200501 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.200637 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.200668 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.200696 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.200714 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:40:50 crc kubenswrapper[4873]: E1201 08:40:50.200771 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:40:58.200719985 +0000 UTC m=+34.102828534 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:40:50 crc kubenswrapper[4873]: E1201 08:40:50.200836 4873 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:40:50 crc kubenswrapper[4873]: E1201 08:40:50.200896 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:40:58.200876719 +0000 UTC m=+34.102985248 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:40:50 crc kubenswrapper[4873]: E1201 08:40:50.200930 4873 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:40:50 crc kubenswrapper[4873]: E1201 08:40:50.200970 4873 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:40:50 crc kubenswrapper[4873]: E1201 08:40:50.200969 4873 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:40:50 crc kubenswrapper[4873]: E1201 08:40:50.200985 4873 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:50 crc kubenswrapper[4873]: E1201 08:40:50.200988 4873 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:40:50 crc kubenswrapper[4873]: E1201 08:40:50.201040 4873 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:40:50 crc kubenswrapper[4873]: E1201 08:40:50.201049 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:40:58.201040563 +0000 UTC m=+34.103149102 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:40:50 crc kubenswrapper[4873]: E1201 08:40:50.201067 4873 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:50 crc kubenswrapper[4873]: E1201 08:40:50.201082 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 08:40:58.201058623 +0000 UTC m=+34.103167162 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:50 crc kubenswrapper[4873]: E1201 08:40:50.201166 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 08:40:58.201151926 +0000 UTC m=+34.103260485 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.253406 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.253459 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.253476 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.253496 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.253511 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:50Z","lastTransitionTime":"2025-12-01T08:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.357410 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.357489 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.357512 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.357544 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.357568 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:50Z","lastTransitionTime":"2025-12-01T08:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.476628 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:40:50 crc kubenswrapper[4873]: E1201 08:40:50.476780 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.477365 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.477431 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:40:50 crc kubenswrapper[4873]: E1201 08:40:50.477552 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:40:50 crc kubenswrapper[4873]: E1201 08:40:50.477615 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.478850 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.478894 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.478909 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.478928 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.478942 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:50Z","lastTransitionTime":"2025-12-01T08:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.581818 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.581901 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.581920 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.581958 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.581972 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:50Z","lastTransitionTime":"2025-12-01T08:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.654866 4873 generic.go:334] "Generic (PLEG): container finished" podID="8570b876-a13b-4b8f-99c4-4900cfe19a21" containerID="0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef" exitCode=0 Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.654942 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" event={"ID":"8570b876-a13b-4b8f-99c4-4900cfe19a21","Type":"ContainerDied","Data":"0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef"} Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.685281 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.685408 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.685516 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.685777 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.685813 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:50Z","lastTransitionTime":"2025-12-01T08:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.698487 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a2348f-009e-4cc1-b591-e1c14ff0faf2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff98ccdc9925e44e62f0d6e826ea7926dfa2a7d83d33e475a092f128b7fe920a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4676d506268107b87ebe2e21ba6e65d959a23a4d6f38bf0de588bd9c5d15e207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4751109230b67ad8948a2de4794571581d662bfd8abc6dc5c198dbc411172860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3acd19dc33ad40d8857d28bc0ccee4c99a6927e92e5029873234ecb5dc6ed18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7242d1034bd1938cb0bd750cab74db52d454ddd8af75f1e9e0a3c007dcdddd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:50Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.722357 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88068a1cf89e2c507b9d93d305db6af5af196bd9eec805f42d91b6fe090aba35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:50Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.743227 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:50Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.765283 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:50Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.795545 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.796057 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.796264 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.796401 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.796554 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:50Z","lastTransitionTime":"2025-12-01T08:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.810852 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a308045b-f54d-42ec-8fdf-5bc3e54ef363\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dxlvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:50Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.841817 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1c8f149-f188-4759-b2dc-9dc15be1bb13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:40:37.024907 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:40:37.026665 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1609525021/tls.crt::/tmp/serving-cert-1609525021/tls.key\\\\\\\"\\\\nI1201 08:40:42.169070 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:40:42.177896 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:40:42.177940 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:40:42.177968 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:40:42.177975 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:40:42.184767 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1201 08:40:42.184781 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1201 08:40:42.184802 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184808 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184814 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:40:42.184819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:40:42.184822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:40:42.184826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1201 08:40:42.187129 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:50Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.869097 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tpwwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13a2246b-93bb-4586-98ee-53fc84aaae02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dfcc763fe6ab69cab09f09c92ad9a85a4ded49c1443904134f78d1ca0092dae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zfr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tpwwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:50Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.887216 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90f08b6fa658ce25fbc25d8d44d956cdb6424a0b047205f19d1a6fd6870b1e90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8d27e3678f1f63f55bd1b76668e328f49368d67d4d7338a6326cb73d13e08f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:50Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.898695 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.898736 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.898745 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.898760 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.898772 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:50Z","lastTransitionTime":"2025-12-01T08:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.899764 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qpr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c42ab88e96b59d9430a69343e06a329447ae0163cb0f5b9ef2bcafa1bf356583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qpr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:50Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.912710 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8570b876-a13b-4b8f-99c4-4900cfe19a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shkn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:50Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.930535 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f7ac0b4-d048-46e4-bcee-4a271503e46d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8208cd5fda4df2e3b5a8bd90a18c8aca62b16cc2c60e06f52411f24f3d60dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba3cfd8a611c687414d7ecd6b02b423b351d7ac2a4d041085d4ca3bb2b0709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ae6257c73f3632c499b5f107178f64a9cac2f4325c57b41aa75d6d4ea14dd12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5088b5bf7bc847f59ca9efbf400855a70b99f1bbbf207eb93576c3b3eabae5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:50Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.944584 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:50Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.959938 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36539974b64191f4ca087bacbd02333f90c89fc8273ab7285e9a1d840155707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:50Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.971374 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef7b114-0e07-402d-a37b-315c36011f4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26e4a6e0a8877a213e2e974d4f7a3ad84938ae1365fa45ca7621bc050470e7e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67d090dbd4e672a6d3e777da244480eaf270518c68f8eaf2292ef7c366b8da4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-scwpp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:50Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:50 crc kubenswrapper[4873]: I1201 08:40:50.981840 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfqcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06182c66-4be6-4e00-9a17-4917badfc997\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa8ef5ef07129ec4a7b4cf4738a01cd6444e9bf548134a9da493c7149301387\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfqcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:50Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.002164 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.002202 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.002213 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.002228 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.002240 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:51Z","lastTransitionTime":"2025-12-01T08:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.105403 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.105468 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.105486 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.105510 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.105527 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:51Z","lastTransitionTime":"2025-12-01T08:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.209466 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.209526 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.209538 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.209561 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.209575 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:51Z","lastTransitionTime":"2025-12-01T08:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.312581 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.312629 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.312639 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.312655 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.312669 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:51Z","lastTransitionTime":"2025-12-01T08:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.414996 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.415062 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.415075 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.415089 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.415099 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:51Z","lastTransitionTime":"2025-12-01T08:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.518904 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.518965 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.518980 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.519001 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.519094 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:51Z","lastTransitionTime":"2025-12-01T08:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.622383 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.622506 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.622922 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.622996 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.623070 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:51Z","lastTransitionTime":"2025-12-01T08:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.666242 4873 generic.go:334] "Generic (PLEG): container finished" podID="8570b876-a13b-4b8f-99c4-4900cfe19a21" containerID="53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a" exitCode=0 Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.666353 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" event={"ID":"8570b876-a13b-4b8f-99c4-4900cfe19a21","Type":"ContainerDied","Data":"53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a"} Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.675634 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" event={"ID":"a308045b-f54d-42ec-8fdf-5bc3e54ef363","Type":"ContainerStarted","Data":"4891456e7daeb7cc841fd0ed34045698e5f1220788c28682baaed9ac3a1e9d36"} Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.676081 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.676235 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.689851 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f7ac0b4-d048-46e4-bcee-4a271503e46d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8208cd5fda4df2e3b5a8bd90a18c8aca62b16cc2c60e06f52411f24f3d60dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba3cfd8a611c687414d7ecd6b02b423b351d7ac2a4d041085d4ca3bb2b0709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ae6257c73f3632c499b5f107178f64a9cac2f4325c57b41aa75d6d4ea14dd12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5088b5bf7bc847f59ca9efbf400855a70b99f1bbbf207eb93576c3b3eabae5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:51Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.714928 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:51Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.726949 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.727639 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.727661 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.727689 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.727710 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:51Z","lastTransitionTime":"2025-12-01T08:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.730461 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36539974b64191f4ca087bacbd02333f90c89fc8273ab7285e9a1d840155707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:51Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.744639 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef7b114-0e07-402d-a37b-315c36011f4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26e4a6e0a8877a213e2e974d4f7a3ad84938ae1365fa45ca7621bc050470e7e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67d090dbd4e672a6d3e777da244480eaf270518c68f8eaf2292ef7c366b8da4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-scwpp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:51Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.756866 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfqcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06182c66-4be6-4e00-9a17-4917badfc997\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa8ef5ef07129ec4a7b4cf4738a01cd6444e9bf548134a9da493c7149301387\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfqcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:51Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.789176 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a2348f-009e-4cc1-b591-e1c14ff0faf2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff98ccdc9925e44e62f0d6e826ea7926dfa2a7d83d33e475a092f128b7fe920a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4676d506268107b87ebe2e21ba6e65d959a23a4d6f38bf0de588bd9c5d15e207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4751109230b67ad8948a2de4794571581d662bfd8abc6dc5c198dbc411172860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3acd19dc33ad40d8857d28bc0ccee4c99a6927e92e5029873234ecb5dc6ed18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7242d1034bd1938cb0bd750cab74db52d454ddd8af75f1e9e0a3c007dcdddd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:51Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.803550 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88068a1cf89e2c507b9d93d305db6af5af196bd9eec805f42d91b6fe090aba35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:51Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.817586 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:51Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.832134 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.832209 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.832231 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.832265 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.832288 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:51Z","lastTransitionTime":"2025-12-01T08:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.841767 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:51Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.864631 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a308045b-f54d-42ec-8fdf-5bc3e54ef363\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dxlvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:51Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.877619 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1c8f149-f188-4759-b2dc-9dc15be1bb13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:40:37.024907 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:40:37.026665 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1609525021/tls.crt::/tmp/serving-cert-1609525021/tls.key\\\\\\\"\\\\nI1201 08:40:42.169070 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:40:42.177896 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:40:42.177940 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:40:42.177968 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:40:42.177975 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:40:42.184767 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1201 08:40:42.184781 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1201 08:40:42.184802 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184808 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184814 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:40:42.184819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:40:42.184822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:40:42.184826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1201 08:40:42.187129 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:51Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.892640 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tpwwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13a2246b-93bb-4586-98ee-53fc84aaae02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dfcc763fe6ab69cab09f09c92ad9a85a4ded49c1443904134f78d1ca0092dae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zfr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tpwwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:51Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.906442 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90f08b6fa658ce25fbc25d8d44d956cdb6424a0b047205f19d1a6fd6870b1e90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8d27e3678f1f63f55bd1b76668e328f49368d67d4d7338a6326cb73d13e08f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:51Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.916237 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qpr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c42ab88e96b59d9430a69343e06a329447ae0163cb0f5b9ef2bcafa1bf356583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qpr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:51Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.934808 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8570b876-a13b-4b8f-99c4-4900cfe19a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shkn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:51Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.935916 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.935959 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.935971 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.935992 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.936038 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:51Z","lastTransitionTime":"2025-12-01T08:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.949350 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90f08b6fa658ce25fbc25d8d44d956cdb6424a0b047205f19d1a6fd6870b1e90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8d27e3678f1f63f55bd1b76668e328f49368d67d4d7338a6326cb73d13e08f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:51Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.960683 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qpr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c42ab88e96b59d9430a69343e06a329447ae0163cb0f5b9ef2bcafa1bf356583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qpr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:51Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.979857 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8570b876-a13b-4b8f-99c4-4900cfe19a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shkn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:51Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:51 crc kubenswrapper[4873]: I1201 08:40:51.993759 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f7ac0b4-d048-46e4-bcee-4a271503e46d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8208cd5fda4df2e3b5a8bd90a18c8aca62b16cc2c60e06f52411f24f3d60dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba3cfd8a611c687414d7ecd6b02b423b351d7ac2a4d041085d4ca3bb2b0709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ae6257c73f3632c499b5f107178f64a9cac2f4325c57b41aa75d6d4ea14dd12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5088b5bf7bc847f59ca9efbf400855a70b99f1bbbf207eb93576c3b3eabae5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:51Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:52 crc kubenswrapper[4873]: I1201 08:40:52.008146 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:52Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:52 crc kubenswrapper[4873]: I1201 08:40:52.019580 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36539974b64191f4ca087bacbd02333f90c89fc8273ab7285e9a1d840155707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:52Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:52 crc kubenswrapper[4873]: I1201 08:40:52.029549 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef7b114-0e07-402d-a37b-315c36011f4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26e4a6e0a8877a213e2e974d4f7a3ad84938ae1365fa45ca7621bc050470e7e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67d090dbd4e672a6d3e777da244480eaf270518c68f8eaf2292ef7c366b8da4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-scwpp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:52Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:52 crc kubenswrapper[4873]: I1201 08:40:52.040443 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfqcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06182c66-4be6-4e00-9a17-4917badfc997\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa8ef5ef07129ec4a7b4cf4738a01cd6444e9bf548134a9da493c7149301387\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfqcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:52Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:52 crc kubenswrapper[4873]: I1201 08:40:52.636408 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:52 crc kubenswrapper[4873]: I1201 08:40:52.636756 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:40:52 crc kubenswrapper[4873]: E1201 08:40:52.637063 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:40:52 crc kubenswrapper[4873]: I1201 08:40:52.636798 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:40:52 crc kubenswrapper[4873]: E1201 08:40:52.637271 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:40:52 crc kubenswrapper[4873]: I1201 08:40:52.637803 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:40:52 crc kubenswrapper[4873]: E1201 08:40:52.637911 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:40:52 crc kubenswrapper[4873]: I1201 08:40:52.638173 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:52 crc kubenswrapper[4873]: I1201 08:40:52.638222 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:52 crc kubenswrapper[4873]: I1201 08:40:52.638239 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:52 crc kubenswrapper[4873]: I1201 08:40:52.638267 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:52 crc kubenswrapper[4873]: I1201 08:40:52.638286 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:52Z","lastTransitionTime":"2025-12-01T08:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:52 crc kubenswrapper[4873]: I1201 08:40:52.687566 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:52 crc kubenswrapper[4873]: I1201 08:40:52.687798 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:52 crc kubenswrapper[4873]: I1201 08:40:52.724192 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a2348f-009e-4cc1-b591-e1c14ff0faf2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff98ccdc9925e44e62f0d6e826ea7926dfa2a7d83d33e475a092f128b7fe920a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4676d506268107b87ebe2e21ba6e65d959a23a4d6f38bf0de588bd9c5d15e207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4751109230b67ad8948a2de4794571581d662bfd8abc6dc5c198dbc411172860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3acd19dc33ad40d8857d28bc0ccee4c99a6927e92e5029873234ecb5dc6ed18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7242d1034bd1938cb0bd750cab74db52d454ddd8af75f1e9e0a3c007dcdddd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:52Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:52 crc kubenswrapper[4873]: I1201 08:40:52.743389 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:52 crc kubenswrapper[4873]: I1201 08:40:52.744636 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:52 crc kubenswrapper[4873]: I1201 08:40:52.744741 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:52 crc kubenswrapper[4873]: I1201 08:40:52.744852 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:52 crc kubenswrapper[4873]: I1201 08:40:52.744937 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:52Z","lastTransitionTime":"2025-12-01T08:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:52 crc kubenswrapper[4873]: I1201 08:40:52.756003 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88068a1cf89e2c507b9d93d305db6af5af196bd9eec805f42d91b6fe090aba35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:52Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:52 crc kubenswrapper[4873]: I1201 08:40:52.773210 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:52Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:52 crc kubenswrapper[4873]: I1201 08:40:52.788374 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:52Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:52 crc kubenswrapper[4873]: I1201 08:40:52.811171 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a308045b-f54d-42ec-8fdf-5bc3e54ef363\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4891456e7daeb7cc841fd0ed34045698e5f1220788c28682baaed9ac3a1e9d36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dxlvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:52Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:52 crc kubenswrapper[4873]: I1201 08:40:52.827469 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1c8f149-f188-4759-b2dc-9dc15be1bb13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:40:37.024907 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:40:37.026665 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1609525021/tls.crt::/tmp/serving-cert-1609525021/tls.key\\\\\\\"\\\\nI1201 08:40:42.169070 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:40:42.177896 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:40:42.177940 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:40:42.177968 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:40:42.177975 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:40:42.184767 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1201 08:40:42.184781 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1201 08:40:42.184802 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184808 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184814 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:40:42.184819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:40:42.184822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:40:42.184826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1201 08:40:42.187129 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:52Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:52 crc kubenswrapper[4873]: I1201 08:40:52.846330 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tpwwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13a2246b-93bb-4586-98ee-53fc84aaae02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dfcc763fe6ab69cab09f09c92ad9a85a4ded49c1443904134f78d1ca0092dae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zfr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tpwwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:52Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:52 crc kubenswrapper[4873]: I1201 08:40:52.848935 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:52 crc kubenswrapper[4873]: I1201 08:40:52.848986 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:52 crc kubenswrapper[4873]: I1201 08:40:52.849000 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:52 crc kubenswrapper[4873]: I1201 08:40:52.849037 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:52 crc kubenswrapper[4873]: I1201 08:40:52.849054 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:52Z","lastTransitionTime":"2025-12-01T08:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:52 crc kubenswrapper[4873]: I1201 08:40:52.862696 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f7ac0b4-d048-46e4-bcee-4a271503e46d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8208cd5fda4df2e3b5a8bd90a18c8aca62b16cc2c60e06f52411f24f3d60dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba3cfd8a611c687414d7ecd6b02b423b351d7ac2a4d041085d4ca3bb2b0709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ae6257c73f3632c499b5f107178f64a9cac2f4325c57b41aa75d6d4ea14dd12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5088b5bf7bc847f59ca9efbf400855a70b99f1bbbf207eb93576c3b3eabae5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:52Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:52 crc kubenswrapper[4873]: I1201 08:40:52.875549 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:52Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:52 crc kubenswrapper[4873]: I1201 08:40:52.889141 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36539974b64191f4ca087bacbd02333f90c89fc8273ab7285e9a1d840155707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:52Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:52 crc kubenswrapper[4873]: I1201 08:40:52.905237 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef7b114-0e07-402d-a37b-315c36011f4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26e4a6e0a8877a213e2e974d4f7a3ad84938ae1365fa45ca7621bc050470e7e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67d090dbd4e672a6d3e777da244480eaf270518c68f8eaf2292ef7c366b8da4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-scwpp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:52Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:52 crc kubenswrapper[4873]: I1201 08:40:52.930106 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfqcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06182c66-4be6-4e00-9a17-4917badfc997\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa8ef5ef07129ec4a7b4cf4738a01cd6444e9bf548134a9da493c7149301387\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfqcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:52Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:52 crc kubenswrapper[4873]: I1201 08:40:52.952410 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:52 crc kubenswrapper[4873]: I1201 08:40:52.952462 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:52 crc kubenswrapper[4873]: I1201 08:40:52.952475 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:52 crc kubenswrapper[4873]: I1201 08:40:52.952494 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:52 crc kubenswrapper[4873]: I1201 08:40:52.952509 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:52Z","lastTransitionTime":"2025-12-01T08:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:52 crc kubenswrapper[4873]: I1201 08:40:52.957363 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a2348f-009e-4cc1-b591-e1c14ff0faf2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff98ccdc9925e44e62f0d6e826ea7926dfa2a7d83d33e475a092f128b7fe920a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4676d506268107b87ebe2e21ba6e65d959a23a4d6f38bf0de588bd9c5d15e207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4751109230b67ad8948a2de4794571581d662bfd8abc6dc5c198dbc411172860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3acd19dc33ad40d8857d28bc0ccee4c99a6927e92e5029873234ecb5dc6ed18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7242d1034bd1938cb0bd750cab74db52d454ddd8af75f1e9e0a3c007dcdddd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:52Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:52 crc kubenswrapper[4873]: I1201 08:40:52.974672 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88068a1cf89e2c507b9d93d305db6af5af196bd9eec805f42d91b6fe090aba35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:52Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:52 crc kubenswrapper[4873]: I1201 08:40:52.995662 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:52Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.011333 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.035634 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a308045b-f54d-42ec-8fdf-5bc3e54ef363\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4891456e7daeb7cc841fd0ed34045698e5f1220788c28682baaed9ac3a1e9d36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dxlvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.054099 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1c8f149-f188-4759-b2dc-9dc15be1bb13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:40:37.024907 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:40:37.026665 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1609525021/tls.crt::/tmp/serving-cert-1609525021/tls.key\\\\\\\"\\\\nI1201 08:40:42.169070 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:40:42.177896 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:40:42.177940 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:40:42.177968 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:40:42.177975 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:40:42.184767 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1201 08:40:42.184781 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1201 08:40:42.184802 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184808 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184814 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:40:42.184819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:40:42.184822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:40:42.184826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1201 08:40:42.187129 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.055622 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.055686 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.055704 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.055722 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.055739 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:53Z","lastTransitionTime":"2025-12-01T08:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.073853 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tpwwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13a2246b-93bb-4586-98ee-53fc84aaae02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dfcc763fe6ab69cab09f09c92ad9a85a4ded49c1443904134f78d1ca0092dae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zfr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tpwwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.087331 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90f08b6fa658ce25fbc25d8d44d956cdb6424a0b047205f19d1a6fd6870b1e90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8d27e3678f1f63f55bd1b76668e328f49368d67d4d7338a6326cb73d13e08f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.097347 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qpr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c42ab88e96b59d9430a69343e06a329447ae0163cb0f5b9ef2bcafa1bf356583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qpr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.118700 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8570b876-a13b-4b8f-99c4-4900cfe19a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shkn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.159644 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.159698 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.159715 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.159737 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.159752 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:53Z","lastTransitionTime":"2025-12-01T08:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.262808 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.262856 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.262871 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.262891 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.262907 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:53Z","lastTransitionTime":"2025-12-01T08:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.366401 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.366458 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.366474 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.366500 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.366540 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:53Z","lastTransitionTime":"2025-12-01T08:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.469410 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.469482 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.469501 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.469527 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.469545 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:53Z","lastTransitionTime":"2025-12-01T08:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.573453 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.573525 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.573544 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.573571 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.573775 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:53Z","lastTransitionTime":"2025-12-01T08:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.677207 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.677272 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.677295 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.677324 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.677346 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:53Z","lastTransitionTime":"2025-12-01T08:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.727620 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" event={"ID":"8570b876-a13b-4b8f-99c4-4900cfe19a21","Type":"ContainerStarted","Data":"0afb7d55f395ee39aa245516631b88aab92399973a628f91172c9e410dcc13dc"} Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.752065 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1c8f149-f188-4759-b2dc-9dc15be1bb13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:40:37.024907 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:40:37.026665 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1609525021/tls.crt::/tmp/serving-cert-1609525021/tls.key\\\\\\\"\\\\nI1201 08:40:42.169070 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:40:42.177896 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:40:42.177940 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:40:42.177968 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:40:42.177975 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:40:42.184767 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1201 08:40:42.184781 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1201 08:40:42.184802 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184808 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184814 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:40:42.184819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:40:42.184822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:40:42.184826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1201 08:40:42.187129 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.772063 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tpwwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13a2246b-93bb-4586-98ee-53fc84aaae02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dfcc763fe6ab69cab09f09c92ad9a85a4ded49c1443904134f78d1ca0092dae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zfr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tpwwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.780113 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.780199 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.780225 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.780259 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.780282 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:53Z","lastTransitionTime":"2025-12-01T08:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.790452 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90f08b6fa658ce25fbc25d8d44d956cdb6424a0b047205f19d1a6fd6870b1e90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8d27e3678f1f63f55bd1b76668e328f49368d67d4d7338a6326cb73d13e08f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.806211 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qpr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c42ab88e96b59d9430a69343e06a329447ae0163cb0f5b9ef2bcafa1bf356583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qpr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.834579 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8570b876-a13b-4b8f-99c4-4900cfe19a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0afb7d55f395ee39aa245516631b88aab92399973a628f91172c9e410dcc13dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shkn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.856728 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f7ac0b4-d048-46e4-bcee-4a271503e46d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8208cd5fda4df2e3b5a8bd90a18c8aca62b16cc2c60e06f52411f24f3d60dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba3cfd8a611c687414d7ecd6b02b423b351d7ac2a4d041085d4ca3bb2b0709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ae6257c73f3632c499b5f107178f64a9cac2f4325c57b41aa75d6d4ea14dd12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5088b5bf7bc847f59ca9efbf400855a70b99f1bbbf207eb93576c3b3eabae5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.877387 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.882993 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.883088 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.883108 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.883137 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.883157 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:53Z","lastTransitionTime":"2025-12-01T08:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.896405 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36539974b64191f4ca087bacbd02333f90c89fc8273ab7285e9a1d840155707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.912457 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef7b114-0e07-402d-a37b-315c36011f4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26e4a6e0a8877a213e2e974d4f7a3ad84938ae1365fa45ca7621bc050470e7e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67d090dbd4e672a6d3e777da244480eaf270518c68f8eaf2292ef7c366b8da4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-scwpp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.928525 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfqcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06182c66-4be6-4e00-9a17-4917badfc997\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa8ef5ef07129ec4a7b4cf4738a01cd6444e9bf548134a9da493c7149301387\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfqcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.962402 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a2348f-009e-4cc1-b591-e1c14ff0faf2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff98ccdc9925e44e62f0d6e826ea7926dfa2a7d83d33e475a092f128b7fe920a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4676d506268107b87ebe2e21ba6e65d959a23a4d6f38bf0de588bd9c5d15e207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4751109230b67ad8948a2de4794571581d662bfd8abc6dc5c198dbc411172860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3acd19dc33ad40d8857d28bc0ccee4c99a6927e92e5029873234ecb5dc6ed18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7242d1034bd1938cb0bd750cab74db52d454ddd8af75f1e9e0a3c007dcdddd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.985323 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88068a1cf89e2c507b9d93d305db6af5af196bd9eec805f42d91b6fe090aba35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.986777 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.986872 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.986893 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.986921 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:53 crc kubenswrapper[4873]: I1201 08:40:53.986944 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:53Z","lastTransitionTime":"2025-12-01T08:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.013459 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:54Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.035109 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:54Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.069771 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a308045b-f54d-42ec-8fdf-5bc3e54ef363\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4891456e7daeb7cc841fd0ed34045698e5f1220788c28682baaed9ac3a1e9d36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dxlvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:54Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.090096 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.090224 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.090251 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.090284 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.090306 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:54Z","lastTransitionTime":"2025-12-01T08:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.192860 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.192908 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.192922 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.192939 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.192954 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:54Z","lastTransitionTime":"2025-12-01T08:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.295404 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.295457 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.295469 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.295485 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.295497 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:54Z","lastTransitionTime":"2025-12-01T08:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.398086 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.398138 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.398151 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.398171 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.398184 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:54Z","lastTransitionTime":"2025-12-01T08:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.429416 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.429553 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.429759 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:40:54 crc kubenswrapper[4873]: E1201 08:40:54.429756 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:40:54 crc kubenswrapper[4873]: E1201 08:40:54.429833 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:40:54 crc kubenswrapper[4873]: E1201 08:40:54.429921 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.446444 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qpr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c42ab88e96b59d9430a69343e06a329447ae0163cb0f5b9ef2bcafa1bf356583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qpr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:54Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.474386 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8570b876-a13b-4b8f-99c4-4900cfe19a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0afb7d55f395ee39aa245516631b88aab92399973a628f91172c9e410dcc13dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shkn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:54Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.499320 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90f08b6fa658ce25fbc25d8d44d956cdb6424a0b047205f19d1a6fd6870b1e90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8d27e3678f1f63f55bd1b76668e328f49368d67d4d7338a6326cb73d13e08f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:54Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.501437 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.501501 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.501518 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.501543 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.501559 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:54Z","lastTransitionTime":"2025-12-01T08:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.515656 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:54Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.526951 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36539974b64191f4ca087bacbd02333f90c89fc8273ab7285e9a1d840155707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:54Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.542253 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef7b114-0e07-402d-a37b-315c36011f4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26e4a6e0a8877a213e2e974d4f7a3ad84938ae1365fa45ca7621bc050470e7e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67d090dbd4e672a6d3e777da244480eaf270518c68f8eaf2292ef7c366b8da4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-scwpp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:54Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.555408 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfqcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06182c66-4be6-4e00-9a17-4917badfc997\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa8ef5ef07129ec4a7b4cf4738a01cd6444e9bf548134a9da493c7149301387\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfqcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:54Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.568548 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f7ac0b4-d048-46e4-bcee-4a271503e46d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8208cd5fda4df2e3b5a8bd90a18c8aca62b16cc2c60e06f52411f24f3d60dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba3cfd8a611c687414d7ecd6b02b423b351d7ac2a4d041085d4ca3bb2b0709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ae6257c73f3632c499b5f107178f64a9cac2f4325c57b41aa75d6d4ea14dd12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5088b5bf7bc847f59ca9efbf400855a70b99f1bbbf207eb93576c3b3eabae5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:54Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.586082 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88068a1cf89e2c507b9d93d305db6af5af196bd9eec805f42d91b6fe090aba35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:54Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.601669 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:54Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.604518 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.604567 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.604580 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.604600 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.604613 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:54Z","lastTransitionTime":"2025-12-01T08:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.616567 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:54Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.635030 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a308045b-f54d-42ec-8fdf-5bc3e54ef363\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4891456e7daeb7cc841fd0ed34045698e5f1220788c28682baaed9ac3a1e9d36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dxlvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:54Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.662064 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a2348f-009e-4cc1-b591-e1c14ff0faf2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff98ccdc9925e44e62f0d6e826ea7926dfa2a7d83d33e475a092f128b7fe920a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4676d506268107b87ebe2e21ba6e65d959a23a4d6f38bf0de588bd9c5d15e207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4751109230b67ad8948a2de4794571581d662bfd8abc6dc5c198dbc411172860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3acd19dc33ad40d8857d28bc0ccee4c99a6927e92e5029873234ecb5dc6ed18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7242d1034bd1938cb0bd750cab74db52d454ddd8af75f1e9e0a3c007dcdddd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:54Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.679127 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1c8f149-f188-4759-b2dc-9dc15be1bb13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:40:37.024907 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:40:37.026665 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1609525021/tls.crt::/tmp/serving-cert-1609525021/tls.key\\\\\\\"\\\\nI1201 08:40:42.169070 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:40:42.177896 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:40:42.177940 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:40:42.177968 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:40:42.177975 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:40:42.184767 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1201 08:40:42.184781 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1201 08:40:42.184802 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184808 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184814 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:40:42.184819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:40:42.184822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:40:42.184826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1201 08:40:42.187129 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:54Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.694327 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tpwwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13a2246b-93bb-4586-98ee-53fc84aaae02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dfcc763fe6ab69cab09f09c92ad9a85a4ded49c1443904134f78d1ca0092dae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zfr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tpwwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:54Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.707182 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.707230 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.707240 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.707253 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.707263 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:54Z","lastTransitionTime":"2025-12-01T08:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.809417 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.809463 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.809474 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.809490 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.809502 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:54Z","lastTransitionTime":"2025-12-01T08:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.913262 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.913313 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.913335 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.913352 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:54 crc kubenswrapper[4873]: I1201 08:40:54.913362 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:54Z","lastTransitionTime":"2025-12-01T08:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.016634 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.016686 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.016701 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.016718 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.016729 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:55Z","lastTransitionTime":"2025-12-01T08:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.119914 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.119959 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.119971 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.119988 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.119999 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:55Z","lastTransitionTime":"2025-12-01T08:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.223463 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.223507 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.223518 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.223533 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.223542 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:55Z","lastTransitionTime":"2025-12-01T08:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.327438 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.327499 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.327517 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.327537 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.327548 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:55Z","lastTransitionTime":"2025-12-01T08:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.430757 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.430810 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.430820 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.430837 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.430849 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:55Z","lastTransitionTime":"2025-12-01T08:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.535154 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.535245 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.535265 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.535290 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.535311 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:55Z","lastTransitionTime":"2025-12-01T08:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.638927 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.638994 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.639050 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.639083 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.639103 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:55Z","lastTransitionTime":"2025-12-01T08:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.660331 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.660393 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.660410 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.660432 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.660449 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:55Z","lastTransitionTime":"2025-12-01T08:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:55 crc kubenswrapper[4873]: E1201 08:40:55.690084 4873 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963d6124-09f6-48b8-b38c-854877aa92e3\\\",\\\"systemUUID\\\":\\\"27c66b0b-2450-400c-b6ea-cdb9cbbf95ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.696782 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.696893 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.696920 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.696951 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.697713 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:55Z","lastTransitionTime":"2025-12-01T08:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:55 crc kubenswrapper[4873]: E1201 08:40:55.718591 4873 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963d6124-09f6-48b8-b38c-854877aa92e3\\\",\\\"systemUUID\\\":\\\"27c66b0b-2450-400c-b6ea-cdb9cbbf95ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.725184 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.725258 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.725278 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.725305 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.725323 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:55Z","lastTransitionTime":"2025-12-01T08:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.739874 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dxlvk_a308045b-f54d-42ec-8fdf-5bc3e54ef363/ovnkube-controller/0.log" Dec 01 08:40:55 crc kubenswrapper[4873]: E1201 08:40:55.744682 4873 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963d6124-09f6-48b8-b38c-854877aa92e3\\\",\\\"systemUUID\\\":\\\"27c66b0b-2450-400c-b6ea-cdb9cbbf95ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.744984 4873 generic.go:334] "Generic (PLEG): container finished" podID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerID="4891456e7daeb7cc841fd0ed34045698e5f1220788c28682baaed9ac3a1e9d36" exitCode=1 Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.745076 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" event={"ID":"a308045b-f54d-42ec-8fdf-5bc3e54ef363","Type":"ContainerDied","Data":"4891456e7daeb7cc841fd0ed34045698e5f1220788c28682baaed9ac3a1e9d36"} Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.746088 4873 scope.go:117] "RemoveContainer" containerID="4891456e7daeb7cc841fd0ed34045698e5f1220788c28682baaed9ac3a1e9d36" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.750108 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.750145 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.750158 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.750178 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.750195 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:55Z","lastTransitionTime":"2025-12-01T08:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:55 crc kubenswrapper[4873]: E1201 08:40:55.771433 4873 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963d6124-09f6-48b8-b38c-854877aa92e3\\\",\\\"systemUUID\\\":\\\"27c66b0b-2450-400c-b6ea-cdb9cbbf95ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.777786 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.777854 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.777881 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.777913 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.777938 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:55Z","lastTransitionTime":"2025-12-01T08:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.785820 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a2348f-009e-4cc1-b591-e1c14ff0faf2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff98ccdc9925e44e62f0d6e826ea7926dfa2a7d83d33e475a092f128b7fe920a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4676d506268107b87ebe2e21ba6e65d959a23a4d6f38bf0de588bd9c5d15e207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4751109230b67ad8948a2de4794571581d662bfd8abc6dc5c198dbc411172860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3acd19dc33ad40d8857d28bc0ccee4c99a6927e92e5029873234ecb5dc6ed18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7242d1034bd1938cb0bd750cab74db52d454ddd8af75f1e9e0a3c007dcdddd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.806663 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88068a1cf89e2c507b9d93d305db6af5af196bd9eec805f42d91b6fe090aba35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4873]: E1201 08:40:55.806979 4873 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963d6124-09f6-48b8-b38c-854877aa92e3\\\",\\\"systemUUID\\\":\\\"27c66b0b-2450-400c-b6ea-cdb9cbbf95ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4873]: E1201 08:40:55.807394 4873 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.811640 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.811695 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.811712 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.811737 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.811755 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:55Z","lastTransitionTime":"2025-12-01T08:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.834069 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.860490 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.894973 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a308045b-f54d-42ec-8fdf-5bc3e54ef363\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4891456e7daeb7cc841fd0ed34045698e5f1220788c28682baaed9ac3a1e9d36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4891456e7daeb7cc841fd0ed34045698e5f1220788c28682baaed9ac3a1e9d36\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:40:54Z\\\",\\\"message\\\":\\\"olicy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1201 08:40:54.703832 6180 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 08:40:54.703905 6180 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 08:40:54.703963 6180 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 08:40:54.704179 6180 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 08:40:54.704427 6180 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1201 08:40:54.704458 6180 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1201 08:40:54.704478 6180 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1201 08:40:54.704490 6180 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1201 08:40:54.704643 6180 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 08:40:54.704671 6180 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1201 08:40:54.704702 6180 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1201 08:40:54.704718 6180 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 08:40:54.704851 6180 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dxlvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.914896 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.914963 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.914988 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.915061 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.915084 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:55Z","lastTransitionTime":"2025-12-01T08:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.926664 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1c8f149-f188-4759-b2dc-9dc15be1bb13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:40:37.024907 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:40:37.026665 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1609525021/tls.crt::/tmp/serving-cert-1609525021/tls.key\\\\\\\"\\\\nI1201 08:40:42.169070 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:40:42.177896 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:40:42.177940 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:40:42.177968 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:40:42.177975 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:40:42.184767 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1201 08:40:42.184781 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1201 08:40:42.184802 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184808 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184814 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:40:42.184819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:40:42.184822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:40:42.184826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1201 08:40:42.187129 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.943708 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tpwwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13a2246b-93bb-4586-98ee-53fc84aaae02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dfcc763fe6ab69cab09f09c92ad9a85a4ded49c1443904134f78d1ca0092dae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zfr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tpwwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4873]: I1201 08:40:55.962847 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90f08b6fa658ce25fbc25d8d44d956cdb6424a0b047205f19d1a6fd6870b1e90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8d27e3678f1f63f55bd1b76668e328f49368d67d4d7338a6326cb73d13e08f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.014703 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qpr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c42ab88e96b59d9430a69343e06a329447ae0163cb0f5b9ef2bcafa1bf356583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qpr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.019525 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.019615 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.019640 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.019674 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.019699 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:56Z","lastTransitionTime":"2025-12-01T08:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.044578 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8570b876-a13b-4b8f-99c4-4900cfe19a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0afb7d55f395ee39aa245516631b88aab92399973a628f91172c9e410dcc13dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shkn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.060317 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f7ac0b4-d048-46e4-bcee-4a271503e46d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8208cd5fda4df2e3b5a8bd90a18c8aca62b16cc2c60e06f52411f24f3d60dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba3cfd8a611c687414d7ecd6b02b423b351d7ac2a4d041085d4ca3bb2b0709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ae6257c73f3632c499b5f107178f64a9cac2f4325c57b41aa75d6d4ea14dd12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5088b5bf7bc847f59ca9efbf400855a70b99f1bbbf207eb93576c3b3eabae5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.077555 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.091862 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36539974b64191f4ca087bacbd02333f90c89fc8273ab7285e9a1d840155707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.104231 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef7b114-0e07-402d-a37b-315c36011f4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26e4a6e0a8877a213e2e974d4f7a3ad84938ae1365fa45ca7621bc050470e7e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67d090dbd4e672a6d3e777da244480eaf270518c68f8eaf2292ef7c366b8da4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-scwpp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.116316 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfqcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06182c66-4be6-4e00-9a17-4917badfc997\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa8ef5ef07129ec4a7b4cf4738a01cd6444e9bf548134a9da493c7149301387\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfqcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.122928 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.122989 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.123004 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.123038 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.123056 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:56Z","lastTransitionTime":"2025-12-01T08:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.227644 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.227737 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.227762 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.227801 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.227821 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:56Z","lastTransitionTime":"2025-12-01T08:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.331740 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.331799 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.331810 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.331831 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.331842 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:56Z","lastTransitionTime":"2025-12-01T08:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.429103 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.429192 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.429298 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:40:56 crc kubenswrapper[4873]: E1201 08:40:56.429344 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:40:56 crc kubenswrapper[4873]: E1201 08:40:56.429483 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:40:56 crc kubenswrapper[4873]: E1201 08:40:56.429627 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.467822 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.467889 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.467914 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.467944 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.467968 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:56Z","lastTransitionTime":"2025-12-01T08:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.572814 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.572861 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.572870 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.572887 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.572899 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:56Z","lastTransitionTime":"2025-12-01T08:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.675806 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.675849 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.675858 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.675873 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.675884 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:56Z","lastTransitionTime":"2025-12-01T08:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.750712 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dxlvk_a308045b-f54d-42ec-8fdf-5bc3e54ef363/ovnkube-controller/0.log" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.753628 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" event={"ID":"a308045b-f54d-42ec-8fdf-5bc3e54ef363","Type":"ContainerStarted","Data":"31ea34e79d7e4ab84a65b4e2e8527dc55075f1f01fecb83ec4c54c0c1c6f1fc1"} Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.754145 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.769298 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1c8f149-f188-4759-b2dc-9dc15be1bb13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:40:37.024907 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:40:37.026665 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1609525021/tls.crt::/tmp/serving-cert-1609525021/tls.key\\\\\\\"\\\\nI1201 08:40:42.169070 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:40:42.177896 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:40:42.177940 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:40:42.177968 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:40:42.177975 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:40:42.184767 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1201 08:40:42.184781 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1201 08:40:42.184802 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184808 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184814 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:40:42.184819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:40:42.184822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:40:42.184826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1201 08:40:42.187129 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.778417 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.778470 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.778482 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.778504 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.778517 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:56Z","lastTransitionTime":"2025-12-01T08:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.782612 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tpwwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13a2246b-93bb-4586-98ee-53fc84aaae02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dfcc763fe6ab69cab09f09c92ad9a85a4ded49c1443904134f78d1ca0092dae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zfr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tpwwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.801816 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90f08b6fa658ce25fbc25d8d44d956cdb6424a0b047205f19d1a6fd6870b1e90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8d27e3678f1f63f55bd1b76668e328f49368d67d4d7338a6326cb73d13e08f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.823980 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qpr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c42ab88e96b59d9430a69343e06a329447ae0163cb0f5b9ef2bcafa1bf356583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qpr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.842090 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8570b876-a13b-4b8f-99c4-4900cfe19a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0afb7d55f395ee39aa245516631b88aab92399973a628f91172c9e410dcc13dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shkn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.856983 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f7ac0b4-d048-46e4-bcee-4a271503e46d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8208cd5fda4df2e3b5a8bd90a18c8aca62b16cc2c60e06f52411f24f3d60dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba3cfd8a611c687414d7ecd6b02b423b351d7ac2a4d041085d4ca3bb2b0709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ae6257c73f3632c499b5f107178f64a9cac2f4325c57b41aa75d6d4ea14dd12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5088b5bf7bc847f59ca9efbf400855a70b99f1bbbf207eb93576c3b3eabae5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.872791 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.881502 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.881547 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.881556 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.881571 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.881584 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:56Z","lastTransitionTime":"2025-12-01T08:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.888577 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36539974b64191f4ca087bacbd02333f90c89fc8273ab7285e9a1d840155707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.902966 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef7b114-0e07-402d-a37b-315c36011f4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26e4a6e0a8877a213e2e974d4f7a3ad84938ae1365fa45ca7621bc050470e7e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67d090dbd4e672a6d3e777da244480eaf270518c68f8eaf2292ef7c366b8da4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-scwpp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.906738 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.918444 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfqcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06182c66-4be6-4e00-9a17-4917badfc997\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa8ef5ef07129ec4a7b4cf4738a01cd6444e9bf548134a9da493c7149301387\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfqcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.938738 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a2348f-009e-4cc1-b591-e1c14ff0faf2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff98ccdc9925e44e62f0d6e826ea7926dfa2a7d83d33e475a092f128b7fe920a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4676d506268107b87ebe2e21ba6e65d959a23a4d6f38bf0de588bd9c5d15e207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4751109230b67ad8948a2de4794571581d662bfd8abc6dc5c198dbc411172860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3acd19dc33ad40d8857d28bc0ccee4c99a6927e92e5029873234ecb5dc6ed18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7242d1034bd1938cb0bd750cab74db52d454ddd8af75f1e9e0a3c007dcdddd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.951920 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88068a1cf89e2c507b9d93d305db6af5af196bd9eec805f42d91b6fe090aba35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.963782 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.974386 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.984417 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.984458 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.984471 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.984494 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.984510 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:56Z","lastTransitionTime":"2025-12-01T08:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:56 crc kubenswrapper[4873]: I1201 08:40:56.992131 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a308045b-f54d-42ec-8fdf-5bc3e54ef363\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ea34e79d7e4ab84a65b4e2e8527dc55075f1f01fecb83ec4c54c0c1c6f1fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4891456e7daeb7cc841fd0ed34045698e5f1220788c28682baaed9ac3a1e9d36\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:40:54Z\\\",\\\"message\\\":\\\"olicy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1201 08:40:54.703832 6180 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 08:40:54.703905 6180 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 08:40:54.703963 6180 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 08:40:54.704179 6180 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 08:40:54.704427 6180 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1201 08:40:54.704458 6180 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1201 08:40:54.704478 6180 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1201 08:40:54.704490 6180 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1201 08:40:54.704643 6180 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 08:40:54.704671 6180 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1201 08:40:54.704702 6180 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1201 08:40:54.704718 6180 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 08:40:54.704851 6180 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dxlvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.006996 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90f08b6fa658ce25fbc25d8d44d956cdb6424a0b047205f19d1a6fd6870b1e90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8d27e3678f1f63f55bd1b76668e328f49368d67d4d7338a6326cb73d13e08f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.017495 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qpr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c42ab88e96b59d9430a69343e06a329447ae0163cb0f5b9ef2bcafa1bf356583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qpr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.032709 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8570b876-a13b-4b8f-99c4-4900cfe19a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0afb7d55f395ee39aa245516631b88aab92399973a628f91172c9e410dcc13dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shkn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.043655 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef7b114-0e07-402d-a37b-315c36011f4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26e4a6e0a8877a213e2e974d4f7a3ad84938ae1365fa45ca7621bc050470e7e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67d090dbd4e672a6d3e777da244480eaf270518c68f8eaf2292ef7c366b8da4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-scwpp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.055309 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfqcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06182c66-4be6-4e00-9a17-4917badfc997\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa8ef5ef07129ec4a7b4cf4738a01cd6444e9bf548134a9da493c7149301387\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfqcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.067506 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f7ac0b4-d048-46e4-bcee-4a271503e46d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8208cd5fda4df2e3b5a8bd90a18c8aca62b16cc2c60e06f52411f24f3d60dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba3cfd8a611c687414d7ecd6b02b423b351d7ac2a4d041085d4ca3bb2b0709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ae6257c73f3632c499b5f107178f64a9cac2f4325c57b41aa75d6d4ea14dd12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5088b5bf7bc847f59ca9efbf400855a70b99f1bbbf207eb93576c3b3eabae5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.081375 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.091445 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.091506 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.091540 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.091562 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.091573 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:57Z","lastTransitionTime":"2025-12-01T08:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.096533 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36539974b64191f4ca087bacbd02333f90c89fc8273ab7285e9a1d840155707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.108219 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.127713 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a308045b-f54d-42ec-8fdf-5bc3e54ef363\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ea34e79d7e4ab84a65b4e2e8527dc55075f1f01fecb83ec4c54c0c1c6f1fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4891456e7daeb7cc841fd0ed34045698e5f1220788c28682baaed9ac3a1e9d36\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:40:54Z\\\",\\\"message\\\":\\\"olicy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1201 08:40:54.703832 6180 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 08:40:54.703905 6180 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 08:40:54.703963 6180 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 08:40:54.704179 6180 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 08:40:54.704427 6180 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1201 08:40:54.704458 6180 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1201 08:40:54.704478 6180 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1201 08:40:54.704490 6180 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1201 08:40:54.704643 6180 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 08:40:54.704671 6180 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1201 08:40:54.704702 6180 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1201 08:40:54.704718 6180 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 08:40:54.704851 6180 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dxlvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.146771 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a2348f-009e-4cc1-b591-e1c14ff0faf2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff98ccdc9925e44e62f0d6e826ea7926dfa2a7d83d33e475a092f128b7fe920a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4676d506268107b87ebe2e21ba6e65d959a23a4d6f38bf0de588bd9c5d15e207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4751109230b67ad8948a2de4794571581d662bfd8abc6dc5c198dbc411172860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3acd19dc33ad40d8857d28bc0ccee4c99a6927e92e5029873234ecb5dc6ed18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7242d1034bd1938cb0bd750cab74db52d454ddd8af75f1e9e0a3c007dcdddd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.160815 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88068a1cf89e2c507b9d93d305db6af5af196bd9eec805f42d91b6fe090aba35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.174519 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.186833 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tpwwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13a2246b-93bb-4586-98ee-53fc84aaae02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dfcc763fe6ab69cab09f09c92ad9a85a4ded49c1443904134f78d1ca0092dae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zfr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tpwwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.194217 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.194259 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.194273 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.194294 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.194308 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:57Z","lastTransitionTime":"2025-12-01T08:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.199839 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1c8f149-f188-4759-b2dc-9dc15be1bb13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:40:37.024907 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:40:37.026665 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1609525021/tls.crt::/tmp/serving-cert-1609525021/tls.key\\\\\\\"\\\\nI1201 08:40:42.169070 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:40:42.177896 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:40:42.177940 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:40:42.177968 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:40:42.177975 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:40:42.184767 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1201 08:40:42.184781 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1201 08:40:42.184802 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184808 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184814 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:40:42.184819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:40:42.184822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:40:42.184826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1201 08:40:42.187129 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.240111 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5smrf"] Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.240612 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5smrf" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.244766 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.245105 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.256386 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f7ac0b4-d048-46e4-bcee-4a271503e46d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8208cd5fda4df2e3b5a8bd90a18c8aca62b16cc2c60e06f52411f24f3d60dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba3cfd8a611c687414d7ecd6b02b423b351d7ac2a4d041085d4ca3bb2b0709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ae6257c73f3632c499b5f107178f64a9cac2f4325c57b41aa75d6d4ea14dd12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5088b5bf7bc847f59ca9efbf400855a70b99f1bbbf207eb93576c3b3eabae5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.269268 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.282401 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36539974b64191f4ca087bacbd02333f90c89fc8273ab7285e9a1d840155707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.285331 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lvr5\" (UniqueName: \"kubernetes.io/projected/8cdde424-6124-432b-9902-cedb9dca26c1-kube-api-access-8lvr5\") pod \"ovnkube-control-plane-749d76644c-5smrf\" (UID: \"8cdde424-6124-432b-9902-cedb9dca26c1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5smrf" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.285390 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8cdde424-6124-432b-9902-cedb9dca26c1-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-5smrf\" (UID: \"8cdde424-6124-432b-9902-cedb9dca26c1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5smrf" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.285429 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8cdde424-6124-432b-9902-cedb9dca26c1-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-5smrf\" (UID: \"8cdde424-6124-432b-9902-cedb9dca26c1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5smrf" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.285460 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8cdde424-6124-432b-9902-cedb9dca26c1-env-overrides\") pod \"ovnkube-control-plane-749d76644c-5smrf\" (UID: \"8cdde424-6124-432b-9902-cedb9dca26c1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5smrf" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.297064 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.297173 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.297188 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.297208 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.297222 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:57Z","lastTransitionTime":"2025-12-01T08:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.297736 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef7b114-0e07-402d-a37b-315c36011f4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26e4a6e0a8877a213e2e974d4f7a3ad84938ae1365fa45ca7621bc050470e7e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67d090dbd4e672a6d3e777da244480eaf270518c68f8eaf2292ef7c366b8da4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-scwpp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.306265 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfqcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06182c66-4be6-4e00-9a17-4917badfc997\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa8ef5ef07129ec4a7b4cf4738a01cd6444e9bf548134a9da493c7149301387\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfqcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.323395 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a2348f-009e-4cc1-b591-e1c14ff0faf2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff98ccdc9925e44e62f0d6e826ea7926dfa2a7d83d33e475a092f128b7fe920a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4676d506268107b87ebe2e21ba6e65d959a23a4d6f38bf0de588bd9c5d15e207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4751109230b67ad8948a2de4794571581d662bfd8abc6dc5c198dbc411172860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3acd19dc33ad40d8857d28bc0ccee4c99a6927e92e5029873234ecb5dc6ed18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7242d1034bd1938cb0bd750cab74db52d454ddd8af75f1e9e0a3c007dcdddd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.336518 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88068a1cf89e2c507b9d93d305db6af5af196bd9eec805f42d91b6fe090aba35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.349065 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.364937 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.386838 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lvr5\" (UniqueName: \"kubernetes.io/projected/8cdde424-6124-432b-9902-cedb9dca26c1-kube-api-access-8lvr5\") pod \"ovnkube-control-plane-749d76644c-5smrf\" (UID: \"8cdde424-6124-432b-9902-cedb9dca26c1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5smrf" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.386890 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8cdde424-6124-432b-9902-cedb9dca26c1-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-5smrf\" (UID: \"8cdde424-6124-432b-9902-cedb9dca26c1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5smrf" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.386911 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8cdde424-6124-432b-9902-cedb9dca26c1-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-5smrf\" (UID: \"8cdde424-6124-432b-9902-cedb9dca26c1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5smrf" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.386946 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8cdde424-6124-432b-9902-cedb9dca26c1-env-overrides\") pod \"ovnkube-control-plane-749d76644c-5smrf\" (UID: \"8cdde424-6124-432b-9902-cedb9dca26c1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5smrf" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.387700 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8cdde424-6124-432b-9902-cedb9dca26c1-env-overrides\") pod \"ovnkube-control-plane-749d76644c-5smrf\" (UID: \"8cdde424-6124-432b-9902-cedb9dca26c1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5smrf" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.388239 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8cdde424-6124-432b-9902-cedb9dca26c1-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-5smrf\" (UID: \"8cdde424-6124-432b-9902-cedb9dca26c1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5smrf" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.393507 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8cdde424-6124-432b-9902-cedb9dca26c1-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-5smrf\" (UID: \"8cdde424-6124-432b-9902-cedb9dca26c1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5smrf" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.398393 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a308045b-f54d-42ec-8fdf-5bc3e54ef363\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ea34e79d7e4ab84a65b4e2e8527dc55075f1f01fecb83ec4c54c0c1c6f1fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4891456e7daeb7cc841fd0ed34045698e5f1220788c28682baaed9ac3a1e9d36\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:40:54Z\\\",\\\"message\\\":\\\"olicy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1201 08:40:54.703832 6180 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 08:40:54.703905 6180 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 08:40:54.703963 6180 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 08:40:54.704179 6180 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 08:40:54.704427 6180 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1201 08:40:54.704458 6180 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1201 08:40:54.704478 6180 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1201 08:40:54.704490 6180 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1201 08:40:54.704643 6180 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 08:40:54.704671 6180 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1201 08:40:54.704702 6180 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1201 08:40:54.704718 6180 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 08:40:54.704851 6180 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dxlvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.400937 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.400980 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.401002 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.401063 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.401090 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:57Z","lastTransitionTime":"2025-12-01T08:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.407838 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lvr5\" (UniqueName: \"kubernetes.io/projected/8cdde424-6124-432b-9902-cedb9dca26c1-kube-api-access-8lvr5\") pod \"ovnkube-control-plane-749d76644c-5smrf\" (UID: \"8cdde424-6124-432b-9902-cedb9dca26c1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5smrf" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.408880 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5smrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdde424-6124-432b-9902-cedb9dca26c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lvr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lvr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5smrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.420487 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1c8f149-f188-4759-b2dc-9dc15be1bb13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:40:37.024907 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:40:37.026665 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1609525021/tls.crt::/tmp/serving-cert-1609525021/tls.key\\\\\\\"\\\\nI1201 08:40:42.169070 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:40:42.177896 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:40:42.177940 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:40:42.177968 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:40:42.177975 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:40:42.184767 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1201 08:40:42.184781 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1201 08:40:42.184802 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184808 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184814 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:40:42.184819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:40:42.184822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:40:42.184826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1201 08:40:42.187129 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.431594 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tpwwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13a2246b-93bb-4586-98ee-53fc84aaae02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dfcc763fe6ab69cab09f09c92ad9a85a4ded49c1443904134f78d1ca0092dae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zfr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tpwwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.443439 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90f08b6fa658ce25fbc25d8d44d956cdb6424a0b047205f19d1a6fd6870b1e90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8d27e3678f1f63f55bd1b76668e328f49368d67d4d7338a6326cb73d13e08f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.455053 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qpr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c42ab88e96b59d9430a69343e06a329447ae0163cb0f5b9ef2bcafa1bf356583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qpr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.471037 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8570b876-a13b-4b8f-99c4-4900cfe19a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0afb7d55f395ee39aa245516631b88aab92399973a628f91172c9e410dcc13dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shkn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.503482 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.503542 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.503560 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.503585 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.503603 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:57Z","lastTransitionTime":"2025-12-01T08:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.556126 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5smrf" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.606091 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.606154 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.606181 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.606210 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.606233 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:57Z","lastTransitionTime":"2025-12-01T08:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.709279 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.709331 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.709378 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.709399 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.709410 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:57Z","lastTransitionTime":"2025-12-01T08:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.764354 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5smrf" event={"ID":"8cdde424-6124-432b-9902-cedb9dca26c1","Type":"ContainerStarted","Data":"1cbb30892e340863ef3bdb5e963f44d6e33b252809b96664c56b0a2277325f12"} Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.765962 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dxlvk_a308045b-f54d-42ec-8fdf-5bc3e54ef363/ovnkube-controller/1.log" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.766507 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dxlvk_a308045b-f54d-42ec-8fdf-5bc3e54ef363/ovnkube-controller/0.log" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.768935 4873 generic.go:334] "Generic (PLEG): container finished" podID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerID="31ea34e79d7e4ab84a65b4e2e8527dc55075f1f01fecb83ec4c54c0c1c6f1fc1" exitCode=1 Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.768965 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" event={"ID":"a308045b-f54d-42ec-8fdf-5bc3e54ef363","Type":"ContainerDied","Data":"31ea34e79d7e4ab84a65b4e2e8527dc55075f1f01fecb83ec4c54c0c1c6f1fc1"} Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.769000 4873 scope.go:117] "RemoveContainer" containerID="4891456e7daeb7cc841fd0ed34045698e5f1220788c28682baaed9ac3a1e9d36" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.769721 4873 scope.go:117] "RemoveContainer" containerID="31ea34e79d7e4ab84a65b4e2e8527dc55075f1f01fecb83ec4c54c0c1c6f1fc1" Dec 01 08:40:57 crc kubenswrapper[4873]: E1201 08:40:57.769865 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dxlvk_openshift-ovn-kubernetes(a308045b-f54d-42ec-8fdf-5bc3e54ef363)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.782603 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1c8f149-f188-4759-b2dc-9dc15be1bb13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:40:37.024907 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:40:37.026665 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1609525021/tls.crt::/tmp/serving-cert-1609525021/tls.key\\\\\\\"\\\\nI1201 08:40:42.169070 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:40:42.177896 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:40:42.177940 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:40:42.177968 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:40:42.177975 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:40:42.184767 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1201 08:40:42.184781 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1201 08:40:42.184802 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184808 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184814 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:40:42.184819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:40:42.184822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:40:42.184826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1201 08:40:42.187129 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.795777 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tpwwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13a2246b-93bb-4586-98ee-53fc84aaae02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dfcc763fe6ab69cab09f09c92ad9a85a4ded49c1443904134f78d1ca0092dae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zfr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tpwwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.812104 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.812148 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.812158 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.812177 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.812187 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:57Z","lastTransitionTime":"2025-12-01T08:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.814178 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90f08b6fa658ce25fbc25d8d44d956cdb6424a0b047205f19d1a6fd6870b1e90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8d27e3678f1f63f55bd1b76668e328f49368d67d4d7338a6326cb73d13e08f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.827091 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qpr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c42ab88e96b59d9430a69343e06a329447ae0163cb0f5b9ef2bcafa1bf356583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qpr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.840736 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8570b876-a13b-4b8f-99c4-4900cfe19a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0afb7d55f395ee39aa245516631b88aab92399973a628f91172c9e410dcc13dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shkn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.853658 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f7ac0b4-d048-46e4-bcee-4a271503e46d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8208cd5fda4df2e3b5a8bd90a18c8aca62b16cc2c60e06f52411f24f3d60dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba3cfd8a611c687414d7ecd6b02b423b351d7ac2a4d041085d4ca3bb2b0709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ae6257c73f3632c499b5f107178f64a9cac2f4325c57b41aa75d6d4ea14dd12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5088b5bf7bc847f59ca9efbf400855a70b99f1bbbf207eb93576c3b3eabae5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.870300 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.887157 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36539974b64191f4ca087bacbd02333f90c89fc8273ab7285e9a1d840155707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.903475 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef7b114-0e07-402d-a37b-315c36011f4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26e4a6e0a8877a213e2e974d4f7a3ad84938ae1365fa45ca7621bc050470e7e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67d090dbd4e672a6d3e777da244480eaf270518c68f8eaf2292ef7c366b8da4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-scwpp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.915387 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.915440 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.915455 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.915478 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.915492 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:57Z","lastTransitionTime":"2025-12-01T08:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.918582 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfqcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06182c66-4be6-4e00-9a17-4917badfc997\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa8ef5ef07129ec4a7b4cf4738a01cd6444e9bf548134a9da493c7149301387\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfqcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.932968 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5smrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdde424-6124-432b-9902-cedb9dca26c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lvr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lvr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5smrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.959371 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a2348f-009e-4cc1-b591-e1c14ff0faf2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff98ccdc9925e44e62f0d6e826ea7926dfa2a7d83d33e475a092f128b7fe920a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4676d506268107b87ebe2e21ba6e65d959a23a4d6f38bf0de588bd9c5d15e207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4751109230b67ad8948a2de4794571581d662bfd8abc6dc5c198dbc411172860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3acd19dc33ad40d8857d28bc0ccee4c99a6927e92e5029873234ecb5dc6ed18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7242d1034bd1938cb0bd750cab74db52d454ddd8af75f1e9e0a3c007dcdddd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.979055 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88068a1cf89e2c507b9d93d305db6af5af196bd9eec805f42d91b6fe090aba35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4873]: I1201 08:40:57.995335 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.010711 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:58Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.018468 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.018507 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.018521 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.018538 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.018549 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:58Z","lastTransitionTime":"2025-12-01T08:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.035967 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a308045b-f54d-42ec-8fdf-5bc3e54ef363\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ea34e79d7e4ab84a65b4e2e8527dc55075f1f01fecb83ec4c54c0c1c6f1fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4891456e7daeb7cc841fd0ed34045698e5f1220788c28682baaed9ac3a1e9d36\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:40:54Z\\\",\\\"message\\\":\\\"olicy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1201 08:40:54.703832 6180 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 08:40:54.703905 6180 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 08:40:54.703963 6180 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 08:40:54.704179 6180 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 08:40:54.704427 6180 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1201 08:40:54.704458 6180 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1201 08:40:54.704478 6180 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1201 08:40:54.704490 6180 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1201 08:40:54.704643 6180 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 08:40:54.704671 6180 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1201 08:40:54.704702 6180 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1201 08:40:54.704718 6180 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 08:40:54.704851 6180 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31ea34e79d7e4ab84a65b4e2e8527dc55075f1f01fecb83ec4c54c0c1c6f1fc1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:40:57Z\\\",\\\"message\\\":\\\"ler 9 for removal\\\\nI1201 08:40:57.266189 6351 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 08:40:57.266200 6351 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 08:40:57.266234 6351 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 08:40:57.266249 6351 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1201 08:40:57.266258 6351 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 08:40:57.266274 6351 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 08:40:57.266284 6351 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 08:40:57.266314 6351 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1201 08:40:57.266318 6351 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 08:40:57.266366 6351 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 08:40:57.266388 6351 factory.go:656] Stopping watch factory\\\\nI1201 08:40:57.266393 6351 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 08:40:57.266407 6351 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1201 08:40:57.266399 6351 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 08:40:57.266449 6351 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dxlvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:58Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.123151 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.123198 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.123211 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.123231 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.123241 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:58Z","lastTransitionTime":"2025-12-01T08:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.226720 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.226778 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.226790 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.226809 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.226820 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:58Z","lastTransitionTime":"2025-12-01T08:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.297732 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.297927 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.297981 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:40:58 crc kubenswrapper[4873]: E1201 08:40:58.298318 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:41:14.29818333 +0000 UTC m=+50.200291879 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.298589 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.298724 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:40:58 crc kubenswrapper[4873]: E1201 08:40:58.298387 4873 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:40:58 crc kubenswrapper[4873]: E1201 08:40:58.298886 4873 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:40:58 crc kubenswrapper[4873]: E1201 08:40:58.298903 4873 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:58 crc kubenswrapper[4873]: E1201 08:40:58.298945 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 08:41:14.29893759 +0000 UTC m=+50.201046129 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:58 crc kubenswrapper[4873]: E1201 08:40:58.298462 4873 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:40:58 crc kubenswrapper[4873]: E1201 08:40:58.299012 4873 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:40:58 crc kubenswrapper[4873]: E1201 08:40:58.299057 4873 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:58 crc kubenswrapper[4873]: E1201 08:40:58.298673 4873 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:40:58 crc kubenswrapper[4873]: E1201 08:40:58.298813 4873 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:40:58 crc kubenswrapper[4873]: E1201 08:40:58.299137 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 08:41:14.299111885 +0000 UTC m=+50.201220454 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:58 crc kubenswrapper[4873]: E1201 08:40:58.299308 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:41:14.299284299 +0000 UTC m=+50.201392878 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:40:58 crc kubenswrapper[4873]: E1201 08:40:58.299333 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:41:14.29932118 +0000 UTC m=+50.201429759 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.330307 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.330353 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.330363 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.330379 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.330392 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:58Z","lastTransitionTime":"2025-12-01T08:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.429739 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.429866 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.429759 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:40:58 crc kubenswrapper[4873]: E1201 08:40:58.429964 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:40:58 crc kubenswrapper[4873]: E1201 08:40:58.430145 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:40:58 crc kubenswrapper[4873]: E1201 08:40:58.430363 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.432862 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.432902 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.432920 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.432944 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.432965 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:58Z","lastTransitionTime":"2025-12-01T08:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.535580 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.535637 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.535655 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.535679 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.535697 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:58Z","lastTransitionTime":"2025-12-01T08:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.638391 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.638467 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.638494 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.638519 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.638543 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:58Z","lastTransitionTime":"2025-12-01T08:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.742851 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.742928 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.742955 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.742983 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.743003 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:58Z","lastTransitionTime":"2025-12-01T08:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.781740 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-9vq2r"] Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.787025 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:40:58 crc kubenswrapper[4873]: E1201 08:40:58.787220 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.793075 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dxlvk_a308045b-f54d-42ec-8fdf-5bc3e54ef363/ovnkube-controller/1.log" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.804919 4873 scope.go:117] "RemoveContainer" containerID="31ea34e79d7e4ab84a65b4e2e8527dc55075f1f01fecb83ec4c54c0c1c6f1fc1" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.805904 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5smrf" event={"ID":"8cdde424-6124-432b-9902-cedb9dca26c1","Type":"ContainerStarted","Data":"3cfdfa53bea0d7e308f7673baf05e104ad7967e0473e36a74a17825433a17c7f"} Dec 01 08:40:58 crc kubenswrapper[4873]: E1201 08:40:58.805926 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dxlvk_openshift-ovn-kubernetes(a308045b-f54d-42ec-8fdf-5bc3e54ef363)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.805962 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5smrf" event={"ID":"8cdde424-6124-432b-9902-cedb9dca26c1","Type":"ContainerStarted","Data":"fd6c098fa4f45096e671331702f2995119b74cc8c87c5fa1b175099070bc9099"} Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.807819 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/08d02e64-a000-4f11-837e-82ad06d925f5-metrics-certs\") pod \"network-metrics-daemon-9vq2r\" (UID: \"08d02e64-a000-4f11-837e-82ad06d925f5\") " pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.807960 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzwvd\" (UniqueName: \"kubernetes.io/projected/08d02e64-a000-4f11-837e-82ad06d925f5-kube-api-access-lzwvd\") pod \"network-metrics-daemon-9vq2r\" (UID: \"08d02e64-a000-4f11-837e-82ad06d925f5\") " pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.819759 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f7ac0b4-d048-46e4-bcee-4a271503e46d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8208cd5fda4df2e3b5a8bd90a18c8aca62b16cc2c60e06f52411f24f3d60dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba3cfd8a611c687414d7ecd6b02b423b351d7ac2a4d041085d4ca3bb2b0709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ae6257c73f3632c499b5f107178f64a9cac2f4325c57b41aa75d6d4ea14dd12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5088b5bf7bc847f59ca9efbf400855a70b99f1bbbf207eb93576c3b3eabae5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:58Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.840467 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:58Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.845540 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.845586 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.845600 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.845619 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.845633 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:58Z","lastTransitionTime":"2025-12-01T08:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.854892 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36539974b64191f4ca087bacbd02333f90c89fc8273ab7285e9a1d840155707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:58Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.867569 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef7b114-0e07-402d-a37b-315c36011f4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26e4a6e0a8877a213e2e974d4f7a3ad84938ae1365fa45ca7621bc050470e7e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67d090dbd4e672a6d3e777da244480eaf270518c68f8eaf2292ef7c366b8da4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-scwpp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:58Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.877901 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfqcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06182c66-4be6-4e00-9a17-4917badfc997\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa8ef5ef07129ec4a7b4cf4738a01cd6444e9bf548134a9da493c7149301387\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfqcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:58Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.888448 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5smrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdde424-6124-432b-9902-cedb9dca26c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lvr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lvr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5smrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:58Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.899132 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9vq2r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08d02e64-a000-4f11-837e-82ad06d925f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzwvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzwvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9vq2r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:58Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.908547 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/08d02e64-a000-4f11-837e-82ad06d925f5-metrics-certs\") pod \"network-metrics-daemon-9vq2r\" (UID: \"08d02e64-a000-4f11-837e-82ad06d925f5\") " pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.908786 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzwvd\" (UniqueName: \"kubernetes.io/projected/08d02e64-a000-4f11-837e-82ad06d925f5-kube-api-access-lzwvd\") pod \"network-metrics-daemon-9vq2r\" (UID: \"08d02e64-a000-4f11-837e-82ad06d925f5\") " pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:40:58 crc kubenswrapper[4873]: E1201 08:40:58.908847 4873 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 08:40:58 crc kubenswrapper[4873]: E1201 08:40:58.908981 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/08d02e64-a000-4f11-837e-82ad06d925f5-metrics-certs podName:08d02e64-a000-4f11-837e-82ad06d925f5 nodeName:}" failed. No retries permitted until 2025-12-01 08:40:59.408949824 +0000 UTC m=+35.311058383 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/08d02e64-a000-4f11-837e-82ad06d925f5-metrics-certs") pod "network-metrics-daemon-9vq2r" (UID: "08d02e64-a000-4f11-837e-82ad06d925f5") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.921978 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a2348f-009e-4cc1-b591-e1c14ff0faf2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff98ccdc9925e44e62f0d6e826ea7926dfa2a7d83d33e475a092f128b7fe920a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4676d506268107b87ebe2e21ba6e65d959a23a4d6f38bf0de588bd9c5d15e207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4751109230b67ad8948a2de4794571581d662bfd8abc6dc5c198dbc411172860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3acd19dc33ad40d8857d28bc0ccee4c99a6927e92e5029873234ecb5dc6ed18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7242d1034bd1938cb0bd750cab74db52d454ddd8af75f1e9e0a3c007dcdddd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:58Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.936211 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzwvd\" (UniqueName: \"kubernetes.io/projected/08d02e64-a000-4f11-837e-82ad06d925f5-kube-api-access-lzwvd\") pod \"network-metrics-daemon-9vq2r\" (UID: \"08d02e64-a000-4f11-837e-82ad06d925f5\") " pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.947218 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88068a1cf89e2c507b9d93d305db6af5af196bd9eec805f42d91b6fe090aba35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:58Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.948978 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.949080 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.949100 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.949127 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.949145 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:58Z","lastTransitionTime":"2025-12-01T08:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.968423 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:58Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:58 crc kubenswrapper[4873]: I1201 08:40:58.988539 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:58Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.022233 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a308045b-f54d-42ec-8fdf-5bc3e54ef363\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ea34e79d7e4ab84a65b4e2e8527dc55075f1f01fecb83ec4c54c0c1c6f1fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4891456e7daeb7cc841fd0ed34045698e5f1220788c28682baaed9ac3a1e9d36\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:40:54Z\\\",\\\"message\\\":\\\"olicy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1201 08:40:54.703832 6180 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 08:40:54.703905 6180 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 08:40:54.703963 6180 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 08:40:54.704179 6180 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 08:40:54.704427 6180 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1201 08:40:54.704458 6180 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1201 08:40:54.704478 6180 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1201 08:40:54.704490 6180 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1201 08:40:54.704643 6180 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 08:40:54.704671 6180 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1201 08:40:54.704702 6180 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1201 08:40:54.704718 6180 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 08:40:54.704851 6180 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31ea34e79d7e4ab84a65b4e2e8527dc55075f1f01fecb83ec4c54c0c1c6f1fc1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:40:57Z\\\",\\\"message\\\":\\\"ler 9 for removal\\\\nI1201 08:40:57.266189 6351 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 08:40:57.266200 6351 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 08:40:57.266234 6351 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 08:40:57.266249 6351 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1201 08:40:57.266258 6351 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 08:40:57.266274 6351 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 08:40:57.266284 6351 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 08:40:57.266314 6351 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1201 08:40:57.266318 6351 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 08:40:57.266366 6351 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 08:40:57.266388 6351 factory.go:656] Stopping watch factory\\\\nI1201 08:40:57.266393 6351 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 08:40:57.266407 6351 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1201 08:40:57.266399 6351 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 08:40:57.266449 6351 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dxlvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:59Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.044576 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1c8f149-f188-4759-b2dc-9dc15be1bb13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:40:37.024907 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:40:37.026665 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1609525021/tls.crt::/tmp/serving-cert-1609525021/tls.key\\\\\\\"\\\\nI1201 08:40:42.169070 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:40:42.177896 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:40:42.177940 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:40:42.177968 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:40:42.177975 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:40:42.184767 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1201 08:40:42.184781 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1201 08:40:42.184802 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184808 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184814 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:40:42.184819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:40:42.184822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:40:42.184826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1201 08:40:42.187129 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:59Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.052109 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.052158 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.052174 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.052206 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.052226 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:59Z","lastTransitionTime":"2025-12-01T08:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.067203 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tpwwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13a2246b-93bb-4586-98ee-53fc84aaae02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dfcc763fe6ab69cab09f09c92ad9a85a4ded49c1443904134f78d1ca0092dae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zfr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tpwwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:59Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.085159 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90f08b6fa658ce25fbc25d8d44d956cdb6424a0b047205f19d1a6fd6870b1e90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8d27e3678f1f63f55bd1b76668e328f49368d67d4d7338a6326cb73d13e08f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:59Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.101489 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qpr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c42ab88e96b59d9430a69343e06a329447ae0163cb0f5b9ef2bcafa1bf356583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qpr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:59Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.124508 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8570b876-a13b-4b8f-99c4-4900cfe19a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0afb7d55f395ee39aa245516631b88aab92399973a628f91172c9e410dcc13dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shkn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:59Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.148721 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90f08b6fa658ce25fbc25d8d44d956cdb6424a0b047205f19d1a6fd6870b1e90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8d27e3678f1f63f55bd1b76668e328f49368d67d4d7338a6326cb73d13e08f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:59Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.156803 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.156882 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.156903 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.156933 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.156954 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:59Z","lastTransitionTime":"2025-12-01T08:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.167229 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qpr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c42ab88e96b59d9430a69343e06a329447ae0163cb0f5b9ef2bcafa1bf356583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qpr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:59Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.191730 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8570b876-a13b-4b8f-99c4-4900cfe19a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0afb7d55f395ee39aa245516631b88aab92399973a628f91172c9e410dcc13dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shkn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:59Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.208401 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef7b114-0e07-402d-a37b-315c36011f4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26e4a6e0a8877a213e2e974d4f7a3ad84938ae1365fa45ca7621bc050470e7e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67d090dbd4e672a6d3e777da244480eaf270518c68f8eaf2292ef7c366b8da4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-scwpp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:59Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.229270 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfqcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06182c66-4be6-4e00-9a17-4917badfc997\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa8ef5ef07129ec4a7b4cf4738a01cd6444e9bf548134a9da493c7149301387\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfqcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:59Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.249695 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f7ac0b4-d048-46e4-bcee-4a271503e46d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8208cd5fda4df2e3b5a8bd90a18c8aca62b16cc2c60e06f52411f24f3d60dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba3cfd8a611c687414d7ecd6b02b423b351d7ac2a4d041085d4ca3bb2b0709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ae6257c73f3632c499b5f107178f64a9cac2f4325c57b41aa75d6d4ea14dd12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5088b5bf7bc847f59ca9efbf400855a70b99f1bbbf207eb93576c3b3eabae5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:59Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.259459 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.259523 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.259541 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.259566 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.259585 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:59Z","lastTransitionTime":"2025-12-01T08:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.268901 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:59Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.288441 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36539974b64191f4ca087bacbd02333f90c89fc8273ab7285e9a1d840155707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:59Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.304609 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:59Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.334868 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a308045b-f54d-42ec-8fdf-5bc3e54ef363\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ea34e79d7e4ab84a65b4e2e8527dc55075f1f01fecb83ec4c54c0c1c6f1fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31ea34e79d7e4ab84a65b4e2e8527dc55075f1f01fecb83ec4c54c0c1c6f1fc1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:40:57Z\\\",\\\"message\\\":\\\"ler 9 for removal\\\\nI1201 08:40:57.266189 6351 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 08:40:57.266200 6351 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 08:40:57.266234 6351 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 08:40:57.266249 6351 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1201 08:40:57.266258 6351 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 08:40:57.266274 6351 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 08:40:57.266284 6351 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 08:40:57.266314 6351 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1201 08:40:57.266318 6351 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 08:40:57.266366 6351 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 08:40:57.266388 6351 factory.go:656] Stopping watch factory\\\\nI1201 08:40:57.266393 6351 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 08:40:57.266407 6351 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1201 08:40:57.266399 6351 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 08:40:57.266449 6351 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dxlvk_openshift-ovn-kubernetes(a308045b-f54d-42ec-8fdf-5bc3e54ef363)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dxlvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:59Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.353414 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5smrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdde424-6124-432b-9902-cedb9dca26c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6c098fa4f45096e671331702f2995119b74cc8c87c5fa1b175099070bc9099\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lvr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cfdfa53bea0d7e308f7673baf05e104ad7967e0473e36a74a17825433a17c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lvr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5smrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:59Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.362623 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.362675 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.362690 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.362709 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.362720 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:59Z","lastTransitionTime":"2025-12-01T08:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.369405 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9vq2r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08d02e64-a000-4f11-837e-82ad06d925f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzwvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzwvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9vq2r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:59Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.395757 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a2348f-009e-4cc1-b591-e1c14ff0faf2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff98ccdc9925e44e62f0d6e826ea7926dfa2a7d83d33e475a092f128b7fe920a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4676d506268107b87ebe2e21ba6e65d959a23a4d6f38bf0de588bd9c5d15e207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4751109230b67ad8948a2de4794571581d662bfd8abc6dc5c198dbc411172860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3acd19dc33ad40d8857d28bc0ccee4c99a6927e92e5029873234ecb5dc6ed18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7242d1034bd1938cb0bd750cab74db52d454ddd8af75f1e9e0a3c007dcdddd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:59Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.414059 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88068a1cf89e2c507b9d93d305db6af5af196bd9eec805f42d91b6fe090aba35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:59Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.416931 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/08d02e64-a000-4f11-837e-82ad06d925f5-metrics-certs\") pod \"network-metrics-daemon-9vq2r\" (UID: \"08d02e64-a000-4f11-837e-82ad06d925f5\") " pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:40:59 crc kubenswrapper[4873]: E1201 08:40:59.417334 4873 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 08:40:59 crc kubenswrapper[4873]: E1201 08:40:59.417506 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/08d02e64-a000-4f11-837e-82ad06d925f5-metrics-certs podName:08d02e64-a000-4f11-837e-82ad06d925f5 nodeName:}" failed. No retries permitted until 2025-12-01 08:41:00.417476028 +0000 UTC m=+36.319584577 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/08d02e64-a000-4f11-837e-82ad06d925f5-metrics-certs") pod "network-metrics-daemon-9vq2r" (UID: "08d02e64-a000-4f11-837e-82ad06d925f5") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.429891 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:59Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.449135 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tpwwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13a2246b-93bb-4586-98ee-53fc84aaae02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dfcc763fe6ab69cab09f09c92ad9a85a4ded49c1443904134f78d1ca0092dae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zfr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tpwwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:59Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.465348 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.465399 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.465412 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.465430 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.465442 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:59Z","lastTransitionTime":"2025-12-01T08:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.466199 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1c8f149-f188-4759-b2dc-9dc15be1bb13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:40:37.024907 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:40:37.026665 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1609525021/tls.crt::/tmp/serving-cert-1609525021/tls.key\\\\\\\"\\\\nI1201 08:40:42.169070 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:40:42.177896 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:40:42.177940 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:40:42.177968 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:40:42.177975 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:40:42.184767 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1201 08:40:42.184781 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1201 08:40:42.184802 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184808 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184814 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:40:42.184819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:40:42.184822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:40:42.184826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1201 08:40:42.187129 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:59Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.568082 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.568130 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.568142 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.568158 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.568171 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:59Z","lastTransitionTime":"2025-12-01T08:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.670452 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.670516 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.670537 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.670562 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.670619 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:59Z","lastTransitionTime":"2025-12-01T08:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.773717 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.773794 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.773822 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.773852 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.773873 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:59Z","lastTransitionTime":"2025-12-01T08:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.877428 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.877514 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.877538 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.877567 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.877590 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:59Z","lastTransitionTime":"2025-12-01T08:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.981518 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.981584 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.981602 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.981628 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:59 crc kubenswrapper[4873]: I1201 08:40:59.981645 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:59Z","lastTransitionTime":"2025-12-01T08:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.084211 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.084277 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.084295 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.084318 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.084336 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:00Z","lastTransitionTime":"2025-12-01T08:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.188387 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.188457 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.188475 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.188504 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.188528 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:00Z","lastTransitionTime":"2025-12-01T08:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.292224 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.292287 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.292305 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.292337 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.292354 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:00Z","lastTransitionTime":"2025-12-01T08:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.395894 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.395963 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.395980 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.396002 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.396083 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:00Z","lastTransitionTime":"2025-12-01T08:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.428076 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/08d02e64-a000-4f11-837e-82ad06d925f5-metrics-certs\") pod \"network-metrics-daemon-9vq2r\" (UID: \"08d02e64-a000-4f11-837e-82ad06d925f5\") " pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:41:00 crc kubenswrapper[4873]: E1201 08:41:00.428278 4873 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 08:41:00 crc kubenswrapper[4873]: E1201 08:41:00.428375 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/08d02e64-a000-4f11-837e-82ad06d925f5-metrics-certs podName:08d02e64-a000-4f11-837e-82ad06d925f5 nodeName:}" failed. No retries permitted until 2025-12-01 08:41:02.428354355 +0000 UTC m=+38.330462904 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/08d02e64-a000-4f11-837e-82ad06d925f5-metrics-certs") pod "network-metrics-daemon-9vq2r" (UID: "08d02e64-a000-4f11-837e-82ad06d925f5") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.428972 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.429001 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.429115 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:41:00 crc kubenswrapper[4873]: E1201 08:41:00.429147 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.429221 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:00 crc kubenswrapper[4873]: E1201 08:41:00.429580 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:41:00 crc kubenswrapper[4873]: E1201 08:41:00.429674 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:00 crc kubenswrapper[4873]: E1201 08:41:00.429747 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.499512 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.499562 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.499572 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.499587 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.499597 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:00Z","lastTransitionTime":"2025-12-01T08:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.603231 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.603294 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.603318 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.603348 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.603369 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:00Z","lastTransitionTime":"2025-12-01T08:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.705991 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.706090 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.706112 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.706140 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.706161 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:00Z","lastTransitionTime":"2025-12-01T08:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.809894 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.809968 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.809986 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.810018 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.810073 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:00Z","lastTransitionTime":"2025-12-01T08:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.913106 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.913186 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.913214 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.913248 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:00 crc kubenswrapper[4873]: I1201 08:41:00.913272 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:00Z","lastTransitionTime":"2025-12-01T08:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.017767 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.017830 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.017850 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.017877 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.017899 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:01Z","lastTransitionTime":"2025-12-01T08:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.122195 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.122289 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.122313 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.122347 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.122370 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:01Z","lastTransitionTime":"2025-12-01T08:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.227300 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.227396 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.227416 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.227442 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.227463 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:01Z","lastTransitionTime":"2025-12-01T08:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.330975 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.331058 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.331074 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.331093 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.331108 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:01Z","lastTransitionTime":"2025-12-01T08:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.434450 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.434539 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.434568 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.434596 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.434620 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:01Z","lastTransitionTime":"2025-12-01T08:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.538516 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.538598 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.538618 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.538647 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.538667 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:01Z","lastTransitionTime":"2025-12-01T08:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.643607 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.643678 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.643699 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.643725 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.643745 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:01Z","lastTransitionTime":"2025-12-01T08:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.747251 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.747330 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.747350 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.747375 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.747392 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:01Z","lastTransitionTime":"2025-12-01T08:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.850202 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.850259 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.850277 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.850299 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.850316 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:01Z","lastTransitionTime":"2025-12-01T08:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.953046 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.953091 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.953102 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.953117 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:01 crc kubenswrapper[4873]: I1201 08:41:01.953128 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:01Z","lastTransitionTime":"2025-12-01T08:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.056158 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.056213 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.056228 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.056248 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.056266 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:02Z","lastTransitionTime":"2025-12-01T08:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.160254 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.160359 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.160377 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.160402 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.160420 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:02Z","lastTransitionTime":"2025-12-01T08:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.263518 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.263574 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.263592 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.263620 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.263636 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:02Z","lastTransitionTime":"2025-12-01T08:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.367965 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.368078 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.368110 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.368139 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.368160 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:02Z","lastTransitionTime":"2025-12-01T08:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.429875 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.429960 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.429993 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.429916 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:02 crc kubenswrapper[4873]: E1201 08:41:02.430110 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:41:02 crc kubenswrapper[4873]: E1201 08:41:02.430217 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:02 crc kubenswrapper[4873]: E1201 08:41:02.430377 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:02 crc kubenswrapper[4873]: E1201 08:41:02.430541 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.453543 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/08d02e64-a000-4f11-837e-82ad06d925f5-metrics-certs\") pod \"network-metrics-daemon-9vq2r\" (UID: \"08d02e64-a000-4f11-837e-82ad06d925f5\") " pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:41:02 crc kubenswrapper[4873]: E1201 08:41:02.453709 4873 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 08:41:02 crc kubenswrapper[4873]: E1201 08:41:02.453772 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/08d02e64-a000-4f11-837e-82ad06d925f5-metrics-certs podName:08d02e64-a000-4f11-837e-82ad06d925f5 nodeName:}" failed. No retries permitted until 2025-12-01 08:41:06.453753193 +0000 UTC m=+42.355861752 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/08d02e64-a000-4f11-837e-82ad06d925f5-metrics-certs") pod "network-metrics-daemon-9vq2r" (UID: "08d02e64-a000-4f11-837e-82ad06d925f5") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.471341 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.471393 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.471406 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.471427 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.471441 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:02Z","lastTransitionTime":"2025-12-01T08:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.574649 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.574716 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.574731 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.574750 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.574764 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:02Z","lastTransitionTime":"2025-12-01T08:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.678916 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.679064 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.679099 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.679135 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.679156 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:02Z","lastTransitionTime":"2025-12-01T08:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.781932 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.781989 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.782002 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.782062 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.782078 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:02Z","lastTransitionTime":"2025-12-01T08:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.885847 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.885923 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.885955 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.885989 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.886010 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:02Z","lastTransitionTime":"2025-12-01T08:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.989216 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.989286 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.989313 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.989343 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:02 crc kubenswrapper[4873]: I1201 08:41:02.989367 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:02Z","lastTransitionTime":"2025-12-01T08:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:03 crc kubenswrapper[4873]: I1201 08:41:03.092735 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:03 crc kubenswrapper[4873]: I1201 08:41:03.092861 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:03 crc kubenswrapper[4873]: I1201 08:41:03.092904 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:03 crc kubenswrapper[4873]: I1201 08:41:03.092940 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:03 crc kubenswrapper[4873]: I1201 08:41:03.092967 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:03Z","lastTransitionTime":"2025-12-01T08:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:03 crc kubenswrapper[4873]: I1201 08:41:03.204825 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:03 crc kubenswrapper[4873]: I1201 08:41:03.204896 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:03 crc kubenswrapper[4873]: I1201 08:41:03.204911 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:03 crc kubenswrapper[4873]: I1201 08:41:03.204936 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:03 crc kubenswrapper[4873]: I1201 08:41:03.204952 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:03Z","lastTransitionTime":"2025-12-01T08:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:03 crc kubenswrapper[4873]: I1201 08:41:03.309056 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:03 crc kubenswrapper[4873]: I1201 08:41:03.309127 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:03 crc kubenswrapper[4873]: I1201 08:41:03.309145 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:03 crc kubenswrapper[4873]: I1201 08:41:03.309172 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:03 crc kubenswrapper[4873]: I1201 08:41:03.309191 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:03Z","lastTransitionTime":"2025-12-01T08:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:03 crc kubenswrapper[4873]: I1201 08:41:03.412545 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:03 crc kubenswrapper[4873]: I1201 08:41:03.412653 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:03 crc kubenswrapper[4873]: I1201 08:41:03.412678 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:03 crc kubenswrapper[4873]: I1201 08:41:03.412716 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:03 crc kubenswrapper[4873]: I1201 08:41:03.412741 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:03Z","lastTransitionTime":"2025-12-01T08:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:03 crc kubenswrapper[4873]: I1201 08:41:03.517252 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:03 crc kubenswrapper[4873]: I1201 08:41:03.517321 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:03 crc kubenswrapper[4873]: I1201 08:41:03.517333 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:03 crc kubenswrapper[4873]: I1201 08:41:03.517362 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:03 crc kubenswrapper[4873]: I1201 08:41:03.517378 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:03Z","lastTransitionTime":"2025-12-01T08:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:03 crc kubenswrapper[4873]: I1201 08:41:03.620997 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:03 crc kubenswrapper[4873]: I1201 08:41:03.621163 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:03 crc kubenswrapper[4873]: I1201 08:41:03.621183 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:03 crc kubenswrapper[4873]: I1201 08:41:03.621210 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:03 crc kubenswrapper[4873]: I1201 08:41:03.621230 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:03Z","lastTransitionTime":"2025-12-01T08:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:03 crc kubenswrapper[4873]: I1201 08:41:03.724360 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:03 crc kubenswrapper[4873]: I1201 08:41:03.724444 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:03 crc kubenswrapper[4873]: I1201 08:41:03.724509 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:03 crc kubenswrapper[4873]: I1201 08:41:03.724542 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:03 crc kubenswrapper[4873]: I1201 08:41:03.724566 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:03Z","lastTransitionTime":"2025-12-01T08:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:03 crc kubenswrapper[4873]: I1201 08:41:03.826824 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:03 crc kubenswrapper[4873]: I1201 08:41:03.826895 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:03 crc kubenswrapper[4873]: I1201 08:41:03.826918 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:03 crc kubenswrapper[4873]: I1201 08:41:03.826947 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:03 crc kubenswrapper[4873]: I1201 08:41:03.826970 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:03Z","lastTransitionTime":"2025-12-01T08:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:03 crc kubenswrapper[4873]: I1201 08:41:03.930801 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:03 crc kubenswrapper[4873]: I1201 08:41:03.930856 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:03 crc kubenswrapper[4873]: I1201 08:41:03.930874 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:03 crc kubenswrapper[4873]: I1201 08:41:03.930904 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:03 crc kubenswrapper[4873]: I1201 08:41:03.930926 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:03Z","lastTransitionTime":"2025-12-01T08:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.034181 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.034235 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.034253 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.034277 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.034294 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:04Z","lastTransitionTime":"2025-12-01T08:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.138003 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.138209 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.138307 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.138396 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.138428 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:04Z","lastTransitionTime":"2025-12-01T08:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.241864 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.242001 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.242063 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.242102 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.242158 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:04Z","lastTransitionTime":"2025-12-01T08:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.346700 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.347249 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.347407 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.347800 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.347967 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:04Z","lastTransitionTime":"2025-12-01T08:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.429277 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.429278 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.429321 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.429359 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:04 crc kubenswrapper[4873]: E1201 08:41:04.429599 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:41:04 crc kubenswrapper[4873]: E1201 08:41:04.429741 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:04 crc kubenswrapper[4873]: E1201 08:41:04.429932 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:04 crc kubenswrapper[4873]: E1201 08:41:04.430071 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.449633 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:04Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.454974 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.455050 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.455063 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.455085 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.455100 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:04Z","lastTransitionTime":"2025-12-01T08:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.473068 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a308045b-f54d-42ec-8fdf-5bc3e54ef363\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ea34e79d7e4ab84a65b4e2e8527dc55075f1f01fecb83ec4c54c0c1c6f1fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31ea34e79d7e4ab84a65b4e2e8527dc55075f1f01fecb83ec4c54c0c1c6f1fc1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:40:57Z\\\",\\\"message\\\":\\\"ler 9 for removal\\\\nI1201 08:40:57.266189 6351 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 08:40:57.266200 6351 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 08:40:57.266234 6351 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 08:40:57.266249 6351 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1201 08:40:57.266258 6351 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 08:40:57.266274 6351 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 08:40:57.266284 6351 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 08:40:57.266314 6351 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1201 08:40:57.266318 6351 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 08:40:57.266366 6351 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 08:40:57.266388 6351 factory.go:656] Stopping watch factory\\\\nI1201 08:40:57.266393 6351 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 08:40:57.266407 6351 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1201 08:40:57.266399 6351 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 08:40:57.266449 6351 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dxlvk_openshift-ovn-kubernetes(a308045b-f54d-42ec-8fdf-5bc3e54ef363)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dxlvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:04Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.488784 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5smrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdde424-6124-432b-9902-cedb9dca26c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6c098fa4f45096e671331702f2995119b74cc8c87c5fa1b175099070bc9099\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lvr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cfdfa53bea0d7e308f7673baf05e104ad7967e0473e36a74a17825433a17c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lvr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5smrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:04Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.502283 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9vq2r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08d02e64-a000-4f11-837e-82ad06d925f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzwvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzwvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9vq2r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:04Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.532163 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a2348f-009e-4cc1-b591-e1c14ff0faf2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff98ccdc9925e44e62f0d6e826ea7926dfa2a7d83d33e475a092f128b7fe920a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4676d506268107b87ebe2e21ba6e65d959a23a4d6f38bf0de588bd9c5d15e207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4751109230b67ad8948a2de4794571581d662bfd8abc6dc5c198dbc411172860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3acd19dc33ad40d8857d28bc0ccee4c99a6927e92e5029873234ecb5dc6ed18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7242d1034bd1938cb0bd750cab74db52d454ddd8af75f1e9e0a3c007dcdddd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:04Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.554427 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88068a1cf89e2c507b9d93d305db6af5af196bd9eec805f42d91b6fe090aba35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:04Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.558313 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.558369 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.558385 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.558404 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.558417 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:04Z","lastTransitionTime":"2025-12-01T08:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.573336 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:04Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.593664 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tpwwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13a2246b-93bb-4586-98ee-53fc84aaae02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dfcc763fe6ab69cab09f09c92ad9a85a4ded49c1443904134f78d1ca0092dae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zfr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tpwwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:04Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.618312 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1c8f149-f188-4759-b2dc-9dc15be1bb13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:40:37.024907 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:40:37.026665 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1609525021/tls.crt::/tmp/serving-cert-1609525021/tls.key\\\\\\\"\\\\nI1201 08:40:42.169070 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:40:42.177896 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:40:42.177940 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:40:42.177968 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:40:42.177975 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:40:42.184767 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1201 08:40:42.184781 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1201 08:40:42.184802 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184808 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184814 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:40:42.184819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:40:42.184822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:40:42.184826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1201 08:40:42.187129 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:04Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.640557 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90f08b6fa658ce25fbc25d8d44d956cdb6424a0b047205f19d1a6fd6870b1e90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8d27e3678f1f63f55bd1b76668e328f49368d67d4d7338a6326cb73d13e08f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:04Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.659054 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qpr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c42ab88e96b59d9430a69343e06a329447ae0163cb0f5b9ef2bcafa1bf356583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qpr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:04Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.661265 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.661322 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.661345 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.661376 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.661400 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:04Z","lastTransitionTime":"2025-12-01T08:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.681290 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8570b876-a13b-4b8f-99c4-4900cfe19a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0afb7d55f395ee39aa245516631b88aab92399973a628f91172c9e410dcc13dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shkn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:04Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.699730 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef7b114-0e07-402d-a37b-315c36011f4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26e4a6e0a8877a213e2e974d4f7a3ad84938ae1365fa45ca7621bc050470e7e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67d090dbd4e672a6d3e777da244480eaf270518c68f8eaf2292ef7c366b8da4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-scwpp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:04Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.715169 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfqcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06182c66-4be6-4e00-9a17-4917badfc997\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa8ef5ef07129ec4a7b4cf4738a01cd6444e9bf548134a9da493c7149301387\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfqcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:04Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.736935 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f7ac0b4-d048-46e4-bcee-4a271503e46d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8208cd5fda4df2e3b5a8bd90a18c8aca62b16cc2c60e06f52411f24f3d60dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba3cfd8a611c687414d7ecd6b02b423b351d7ac2a4d041085d4ca3bb2b0709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ae6257c73f3632c499b5f107178f64a9cac2f4325c57b41aa75d6d4ea14dd12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5088b5bf7bc847f59ca9efbf400855a70b99f1bbbf207eb93576c3b3eabae5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:04Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.755637 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:04Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.763798 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.763844 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.763857 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.763874 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.763920 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:04Z","lastTransitionTime":"2025-12-01T08:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.774954 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36539974b64191f4ca087bacbd02333f90c89fc8273ab7285e9a1d840155707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:04Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.867338 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.867394 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.867409 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.867435 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.867451 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:04Z","lastTransitionTime":"2025-12-01T08:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.970899 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.970952 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.970970 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.970995 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:04 crc kubenswrapper[4873]: I1201 08:41:04.971012 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:04Z","lastTransitionTime":"2025-12-01T08:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:05 crc kubenswrapper[4873]: I1201 08:41:05.074566 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:05 crc kubenswrapper[4873]: I1201 08:41:05.074623 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:05 crc kubenswrapper[4873]: I1201 08:41:05.074638 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:05 crc kubenswrapper[4873]: I1201 08:41:05.074662 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:05 crc kubenswrapper[4873]: I1201 08:41:05.074676 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:05Z","lastTransitionTime":"2025-12-01T08:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:05 crc kubenswrapper[4873]: I1201 08:41:05.178352 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:05 crc kubenswrapper[4873]: I1201 08:41:05.178426 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:05 crc kubenswrapper[4873]: I1201 08:41:05.178449 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:05 crc kubenswrapper[4873]: I1201 08:41:05.178473 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:05 crc kubenswrapper[4873]: I1201 08:41:05.178490 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:05Z","lastTransitionTime":"2025-12-01T08:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:05 crc kubenswrapper[4873]: I1201 08:41:05.281253 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:05 crc kubenswrapper[4873]: I1201 08:41:05.281335 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:05 crc kubenswrapper[4873]: I1201 08:41:05.281359 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:05 crc kubenswrapper[4873]: I1201 08:41:05.281388 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:05 crc kubenswrapper[4873]: I1201 08:41:05.281410 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:05Z","lastTransitionTime":"2025-12-01T08:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:05 crc kubenswrapper[4873]: I1201 08:41:05.384762 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:05 crc kubenswrapper[4873]: I1201 08:41:05.384842 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:05 crc kubenswrapper[4873]: I1201 08:41:05.384866 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:05 crc kubenswrapper[4873]: I1201 08:41:05.384895 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:05 crc kubenswrapper[4873]: I1201 08:41:05.384916 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:05Z","lastTransitionTime":"2025-12-01T08:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:05 crc kubenswrapper[4873]: I1201 08:41:05.487827 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:05 crc kubenswrapper[4873]: I1201 08:41:05.487884 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:05 crc kubenswrapper[4873]: I1201 08:41:05.487897 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:05 crc kubenswrapper[4873]: I1201 08:41:05.487920 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:05 crc kubenswrapper[4873]: I1201 08:41:05.487933 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:05Z","lastTransitionTime":"2025-12-01T08:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:05 crc kubenswrapper[4873]: I1201 08:41:05.590875 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:05 crc kubenswrapper[4873]: I1201 08:41:05.590954 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:05 crc kubenswrapper[4873]: I1201 08:41:05.590976 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:05 crc kubenswrapper[4873]: I1201 08:41:05.591008 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:05 crc kubenswrapper[4873]: I1201 08:41:05.591073 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:05Z","lastTransitionTime":"2025-12-01T08:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:05 crc kubenswrapper[4873]: I1201 08:41:05.694069 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:05 crc kubenswrapper[4873]: I1201 08:41:05.694142 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:05 crc kubenswrapper[4873]: I1201 08:41:05.694219 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:05 crc kubenswrapper[4873]: I1201 08:41:05.694299 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:05 crc kubenswrapper[4873]: I1201 08:41:05.694328 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:05Z","lastTransitionTime":"2025-12-01T08:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:05 crc kubenswrapper[4873]: I1201 08:41:05.797817 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:05 crc kubenswrapper[4873]: I1201 08:41:05.797898 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:05 crc kubenswrapper[4873]: I1201 08:41:05.797921 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:05 crc kubenswrapper[4873]: I1201 08:41:05.797955 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:05 crc kubenswrapper[4873]: I1201 08:41:05.797978 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:05Z","lastTransitionTime":"2025-12-01T08:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:05 crc kubenswrapper[4873]: I1201 08:41:05.901893 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:05 crc kubenswrapper[4873]: I1201 08:41:05.901941 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:05 crc kubenswrapper[4873]: I1201 08:41:05.901951 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:05 crc kubenswrapper[4873]: I1201 08:41:05.901968 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:05 crc kubenswrapper[4873]: I1201 08:41:05.901978 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:05Z","lastTransitionTime":"2025-12-01T08:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.005235 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.005306 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.005324 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.005348 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.005374 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:06Z","lastTransitionTime":"2025-12-01T08:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.029125 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.029197 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.029218 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.029243 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.029261 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:06Z","lastTransitionTime":"2025-12-01T08:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:06 crc kubenswrapper[4873]: E1201 08:41:06.051643 4873 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963d6124-09f6-48b8-b38c-854877aa92e3\\\",\\\"systemUUID\\\":\\\"27c66b0b-2450-400c-b6ea-cdb9cbbf95ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.058115 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.058198 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.058225 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.058257 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.058276 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:06Z","lastTransitionTime":"2025-12-01T08:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:06 crc kubenswrapper[4873]: E1201 08:41:06.081348 4873 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963d6124-09f6-48b8-b38c-854877aa92e3\\\",\\\"systemUUID\\\":\\\"27c66b0b-2450-400c-b6ea-cdb9cbbf95ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.087810 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.087897 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.087953 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.087987 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.088070 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:06Z","lastTransitionTime":"2025-12-01T08:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:06 crc kubenswrapper[4873]: E1201 08:41:06.106735 4873 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963d6124-09f6-48b8-b38c-854877aa92e3\\\",\\\"systemUUID\\\":\\\"27c66b0b-2450-400c-b6ea-cdb9cbbf95ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.112478 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.112535 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.112556 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.112581 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.112645 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:06Z","lastTransitionTime":"2025-12-01T08:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:06 crc kubenswrapper[4873]: E1201 08:41:06.132490 4873 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963d6124-09f6-48b8-b38c-854877aa92e3\\\",\\\"systemUUID\\\":\\\"27c66b0b-2450-400c-b6ea-cdb9cbbf95ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.137265 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.137303 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.137311 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.137327 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.137336 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:06Z","lastTransitionTime":"2025-12-01T08:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:06 crc kubenswrapper[4873]: E1201 08:41:06.155456 4873 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963d6124-09f6-48b8-b38c-854877aa92e3\\\",\\\"systemUUID\\\":\\\"27c66b0b-2450-400c-b6ea-cdb9cbbf95ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4873]: E1201 08:41:06.155692 4873 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.158402 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.158441 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.158459 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.158484 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.158503 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:06Z","lastTransitionTime":"2025-12-01T08:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.261761 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.261829 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.261849 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.261873 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.261894 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:06Z","lastTransitionTime":"2025-12-01T08:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.365123 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.365201 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.365223 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.365254 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.365278 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:06Z","lastTransitionTime":"2025-12-01T08:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.428978 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.429113 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.428977 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:41:06 crc kubenswrapper[4873]: E1201 08:41:06.429190 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.429336 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:06 crc kubenswrapper[4873]: E1201 08:41:06.429328 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:06 crc kubenswrapper[4873]: E1201 08:41:06.429580 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:41:06 crc kubenswrapper[4873]: E1201 08:41:06.429972 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.468897 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.468962 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.468982 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.469008 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.469127 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:06Z","lastTransitionTime":"2025-12-01T08:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.515787 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/08d02e64-a000-4f11-837e-82ad06d925f5-metrics-certs\") pod \"network-metrics-daemon-9vq2r\" (UID: \"08d02e64-a000-4f11-837e-82ad06d925f5\") " pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:41:06 crc kubenswrapper[4873]: E1201 08:41:06.515976 4873 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 08:41:06 crc kubenswrapper[4873]: E1201 08:41:06.516092 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/08d02e64-a000-4f11-837e-82ad06d925f5-metrics-certs podName:08d02e64-a000-4f11-837e-82ad06d925f5 nodeName:}" failed. No retries permitted until 2025-12-01 08:41:14.516066352 +0000 UTC m=+50.418174921 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/08d02e64-a000-4f11-837e-82ad06d925f5-metrics-certs") pod "network-metrics-daemon-9vq2r" (UID: "08d02e64-a000-4f11-837e-82ad06d925f5") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.572300 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.572363 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.572385 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.572411 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.572430 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:06Z","lastTransitionTime":"2025-12-01T08:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.676068 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.676134 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.676148 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.676169 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.676182 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:06Z","lastTransitionTime":"2025-12-01T08:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.779394 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.779452 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.779465 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.779488 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.779501 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:06Z","lastTransitionTime":"2025-12-01T08:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.882755 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.882831 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.882845 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.882873 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.882888 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:06Z","lastTransitionTime":"2025-12-01T08:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.986667 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.986811 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.986845 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.986880 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:06 crc kubenswrapper[4873]: I1201 08:41:06.986903 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:06Z","lastTransitionTime":"2025-12-01T08:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:07 crc kubenswrapper[4873]: I1201 08:41:07.090816 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:07 crc kubenswrapper[4873]: I1201 08:41:07.090882 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:07 crc kubenswrapper[4873]: I1201 08:41:07.090902 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:07 crc kubenswrapper[4873]: I1201 08:41:07.090929 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:07 crc kubenswrapper[4873]: I1201 08:41:07.090948 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:07Z","lastTransitionTime":"2025-12-01T08:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:07 crc kubenswrapper[4873]: I1201 08:41:07.194552 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:07 crc kubenswrapper[4873]: I1201 08:41:07.194609 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:07 crc kubenswrapper[4873]: I1201 08:41:07.194622 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:07 crc kubenswrapper[4873]: I1201 08:41:07.194647 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:07 crc kubenswrapper[4873]: I1201 08:41:07.194661 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:07Z","lastTransitionTime":"2025-12-01T08:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:07 crc kubenswrapper[4873]: I1201 08:41:07.297424 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:07 crc kubenswrapper[4873]: I1201 08:41:07.297492 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:07 crc kubenswrapper[4873]: I1201 08:41:07.297509 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:07 crc kubenswrapper[4873]: I1201 08:41:07.297538 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:07 crc kubenswrapper[4873]: I1201 08:41:07.297554 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:07Z","lastTransitionTime":"2025-12-01T08:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:07 crc kubenswrapper[4873]: I1201 08:41:07.400238 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:07 crc kubenswrapper[4873]: I1201 08:41:07.400287 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:07 crc kubenswrapper[4873]: I1201 08:41:07.400313 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:07 crc kubenswrapper[4873]: I1201 08:41:07.400339 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:07 crc kubenswrapper[4873]: I1201 08:41:07.400357 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:07Z","lastTransitionTime":"2025-12-01T08:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:07 crc kubenswrapper[4873]: I1201 08:41:07.503403 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:07 crc kubenswrapper[4873]: I1201 08:41:07.503475 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:07 crc kubenswrapper[4873]: I1201 08:41:07.503488 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:07 crc kubenswrapper[4873]: I1201 08:41:07.503513 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:07 crc kubenswrapper[4873]: I1201 08:41:07.503528 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:07Z","lastTransitionTime":"2025-12-01T08:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:07 crc kubenswrapper[4873]: I1201 08:41:07.607354 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:07 crc kubenswrapper[4873]: I1201 08:41:07.607399 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:07 crc kubenswrapper[4873]: I1201 08:41:07.607410 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:07 crc kubenswrapper[4873]: I1201 08:41:07.607450 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:07 crc kubenswrapper[4873]: I1201 08:41:07.607463 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:07Z","lastTransitionTime":"2025-12-01T08:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:07 crc kubenswrapper[4873]: I1201 08:41:07.714162 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:07 crc kubenswrapper[4873]: I1201 08:41:07.714286 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:07 crc kubenswrapper[4873]: I1201 08:41:07.714309 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:07 crc kubenswrapper[4873]: I1201 08:41:07.714372 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:07 crc kubenswrapper[4873]: I1201 08:41:07.714395 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:07Z","lastTransitionTime":"2025-12-01T08:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:07 crc kubenswrapper[4873]: I1201 08:41:07.824575 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:07 crc kubenswrapper[4873]: I1201 08:41:07.824649 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:07 crc kubenswrapper[4873]: I1201 08:41:07.824676 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:07 crc kubenswrapper[4873]: I1201 08:41:07.824706 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:07 crc kubenswrapper[4873]: I1201 08:41:07.824733 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:07Z","lastTransitionTime":"2025-12-01T08:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:07 crc kubenswrapper[4873]: I1201 08:41:07.928060 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:07 crc kubenswrapper[4873]: I1201 08:41:07.928135 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:07 crc kubenswrapper[4873]: I1201 08:41:07.928155 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:07 crc kubenswrapper[4873]: I1201 08:41:07.928184 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:07 crc kubenswrapper[4873]: I1201 08:41:07.928207 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:07Z","lastTransitionTime":"2025-12-01T08:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.031797 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.031857 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.031876 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.031900 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.031920 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:08Z","lastTransitionTime":"2025-12-01T08:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.135616 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.135682 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.135706 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.135735 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.135819 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:08Z","lastTransitionTime":"2025-12-01T08:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.238357 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.238455 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.238487 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.238518 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.238542 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:08Z","lastTransitionTime":"2025-12-01T08:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.341571 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.341612 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.341625 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.341641 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.341655 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:08Z","lastTransitionTime":"2025-12-01T08:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.429332 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.429381 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.429388 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:41:08 crc kubenswrapper[4873]: E1201 08:41:08.429525 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:08 crc kubenswrapper[4873]: E1201 08:41:08.429702 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.429772 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:08 crc kubenswrapper[4873]: E1201 08:41:08.429856 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:08 crc kubenswrapper[4873]: E1201 08:41:08.429955 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.444517 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.444568 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.444581 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.444601 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.444614 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:08Z","lastTransitionTime":"2025-12-01T08:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.548367 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.548430 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.548440 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.548456 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.548469 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:08Z","lastTransitionTime":"2025-12-01T08:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.652298 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.652441 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.652466 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.652502 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.652525 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:08Z","lastTransitionTime":"2025-12-01T08:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.756189 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.756241 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.756258 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.756281 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.756300 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:08Z","lastTransitionTime":"2025-12-01T08:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.858717 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.859098 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.859166 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.859238 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.859384 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:08Z","lastTransitionTime":"2025-12-01T08:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.962190 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.962260 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.962278 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.962308 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:08 crc kubenswrapper[4873]: I1201 08:41:08.962327 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:08Z","lastTransitionTime":"2025-12-01T08:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.066418 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.066551 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.066571 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.066601 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.066621 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:09Z","lastTransitionTime":"2025-12-01T08:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.169686 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.169731 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.169744 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.169761 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.169774 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:09Z","lastTransitionTime":"2025-12-01T08:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.273451 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.273537 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.273550 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.273576 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.273591 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:09Z","lastTransitionTime":"2025-12-01T08:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.376799 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.376877 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.376891 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.376910 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.376922 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:09Z","lastTransitionTime":"2025-12-01T08:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.430168 4873 scope.go:117] "RemoveContainer" containerID="31ea34e79d7e4ab84a65b4e2e8527dc55075f1f01fecb83ec4c54c0c1c6f1fc1" Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.480897 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.481403 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.481416 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.481461 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.481476 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:09Z","lastTransitionTime":"2025-12-01T08:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.584881 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.584954 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.584971 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.584996 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.585041 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:09Z","lastTransitionTime":"2025-12-01T08:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.693334 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.693389 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.693404 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.693424 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.693442 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:09Z","lastTransitionTime":"2025-12-01T08:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.798957 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.799038 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.799051 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.799069 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.799115 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:09Z","lastTransitionTime":"2025-12-01T08:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.847663 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dxlvk_a308045b-f54d-42ec-8fdf-5bc3e54ef363/ovnkube-controller/1.log" Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.850547 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" event={"ID":"a308045b-f54d-42ec-8fdf-5bc3e54ef363","Type":"ContainerStarted","Data":"d41610d208e2751d8f43b01b041454558f7764717c0855470dd5680a61a34e6f"} Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.851175 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.864607 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1c8f149-f188-4759-b2dc-9dc15be1bb13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:40:37.024907 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:40:37.026665 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1609525021/tls.crt::/tmp/serving-cert-1609525021/tls.key\\\\\\\"\\\\nI1201 08:40:42.169070 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:40:42.177896 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:40:42.177940 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:40:42.177968 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:40:42.177975 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:40:42.184767 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1201 08:40:42.184781 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1201 08:40:42.184802 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184808 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184814 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:40:42.184819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:40:42.184822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:40:42.184826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1201 08:40:42.187129 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:09Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.879636 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tpwwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13a2246b-93bb-4586-98ee-53fc84aaae02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dfcc763fe6ab69cab09f09c92ad9a85a4ded49c1443904134f78d1ca0092dae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zfr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tpwwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:09Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.891677 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90f08b6fa658ce25fbc25d8d44d956cdb6424a0b047205f19d1a6fd6870b1e90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8d27e3678f1f63f55bd1b76668e328f49368d67d4d7338a6326cb73d13e08f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:09Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.901797 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.901851 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.901864 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.901885 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.901897 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:09Z","lastTransitionTime":"2025-12-01T08:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.902254 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qpr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c42ab88e96b59d9430a69343e06a329447ae0163cb0f5b9ef2bcafa1bf356583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qpr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:09Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.917945 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8570b876-a13b-4b8f-99c4-4900cfe19a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0afb7d55f395ee39aa245516631b88aab92399973a628f91172c9e410dcc13dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shkn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:09Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.930685 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfqcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06182c66-4be6-4e00-9a17-4917badfc997\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa8ef5ef07129ec4a7b4cf4738a01cd6444e9bf548134a9da493c7149301387\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfqcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:09Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.947947 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f7ac0b4-d048-46e4-bcee-4a271503e46d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8208cd5fda4df2e3b5a8bd90a18c8aca62b16cc2c60e06f52411f24f3d60dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba3cfd8a611c687414d7ecd6b02b423b351d7ac2a4d041085d4ca3bb2b0709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ae6257c73f3632c499b5f107178f64a9cac2f4325c57b41aa75d6d4ea14dd12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5088b5bf7bc847f59ca9efbf400855a70b99f1bbbf207eb93576c3b3eabae5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:09Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.965235 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:09Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:09 crc kubenswrapper[4873]: I1201 08:41:09.983140 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36539974b64191f4ca087bacbd02333f90c89fc8273ab7285e9a1d840155707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:09Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.001299 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef7b114-0e07-402d-a37b-315c36011f4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26e4a6e0a8877a213e2e974d4f7a3ad84938ae1365fa45ca7621bc050470e7e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67d090dbd4e672a6d3e777da244480eaf270518c68f8eaf2292ef7c366b8da4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-scwpp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:09Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.005350 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.005393 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.005406 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.005421 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.005432 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:10Z","lastTransitionTime":"2025-12-01T08:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.021167 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a308045b-f54d-42ec-8fdf-5bc3e54ef363\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d41610d208e2751d8f43b01b041454558f7764717c0855470dd5680a61a34e6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31ea34e79d7e4ab84a65b4e2e8527dc55075f1f01fecb83ec4c54c0c1c6f1fc1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:40:57Z\\\",\\\"message\\\":\\\"ler 9 for removal\\\\nI1201 08:40:57.266189 6351 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 08:40:57.266200 6351 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 08:40:57.266234 6351 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 08:40:57.266249 6351 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1201 08:40:57.266258 6351 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 08:40:57.266274 6351 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 08:40:57.266284 6351 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 08:40:57.266314 6351 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1201 08:40:57.266318 6351 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 08:40:57.266366 6351 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 08:40:57.266388 6351 factory.go:656] Stopping watch factory\\\\nI1201 08:40:57.266393 6351 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 08:40:57.266407 6351 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1201 08:40:57.266399 6351 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 08:40:57.266449 6351 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:41:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dxlvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:10Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.031600 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5smrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdde424-6124-432b-9902-cedb9dca26c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6c098fa4f45096e671331702f2995119b74cc8c87c5fa1b175099070bc9099\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lvr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cfdfa53bea0d7e308f7673baf05e104ad7967e0473e36a74a17825433a17c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lvr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5smrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:10Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.041939 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9vq2r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08d02e64-a000-4f11-837e-82ad06d925f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzwvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzwvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9vq2r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:10Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.059115 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a2348f-009e-4cc1-b591-e1c14ff0faf2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff98ccdc9925e44e62f0d6e826ea7926dfa2a7d83d33e475a092f128b7fe920a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4676d506268107b87ebe2e21ba6e65d959a23a4d6f38bf0de588bd9c5d15e207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4751109230b67ad8948a2de4794571581d662bfd8abc6dc5c198dbc411172860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3acd19dc33ad40d8857d28bc0ccee4c99a6927e92e5029873234ecb5dc6ed18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7242d1034bd1938cb0bd750cab74db52d454ddd8af75f1e9e0a3c007dcdddd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:10Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.072181 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88068a1cf89e2c507b9d93d305db6af5af196bd9eec805f42d91b6fe090aba35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:10Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.085932 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:10Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.098512 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:10Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.108627 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.108667 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.108676 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.108692 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.108703 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:10Z","lastTransitionTime":"2025-12-01T08:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.212225 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.212264 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.212274 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.212289 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.212298 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:10Z","lastTransitionTime":"2025-12-01T08:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.315775 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.315824 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.315834 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.315852 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.315863 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:10Z","lastTransitionTime":"2025-12-01T08:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.418558 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.418622 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.418638 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.418661 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.418679 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:10Z","lastTransitionTime":"2025-12-01T08:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.430082 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.430127 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.430162 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.430080 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:10 crc kubenswrapper[4873]: E1201 08:41:10.430275 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:41:10 crc kubenswrapper[4873]: E1201 08:41:10.430417 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:10 crc kubenswrapper[4873]: E1201 08:41:10.430556 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:10 crc kubenswrapper[4873]: E1201 08:41:10.430687 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.522175 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.522231 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.522244 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.522264 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.522279 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:10Z","lastTransitionTime":"2025-12-01T08:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.625772 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.625812 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.625823 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.625838 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.625849 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:10Z","lastTransitionTime":"2025-12-01T08:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.729823 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.729951 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.729992 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.730062 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.730090 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:10Z","lastTransitionTime":"2025-12-01T08:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.833682 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.833763 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.833789 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.833821 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.833844 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:10Z","lastTransitionTime":"2025-12-01T08:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.857172 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dxlvk_a308045b-f54d-42ec-8fdf-5bc3e54ef363/ovnkube-controller/2.log" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.857819 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dxlvk_a308045b-f54d-42ec-8fdf-5bc3e54ef363/ovnkube-controller/1.log" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.861803 4873 generic.go:334] "Generic (PLEG): container finished" podID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerID="d41610d208e2751d8f43b01b041454558f7764717c0855470dd5680a61a34e6f" exitCode=1 Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.861870 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" event={"ID":"a308045b-f54d-42ec-8fdf-5bc3e54ef363","Type":"ContainerDied","Data":"d41610d208e2751d8f43b01b041454558f7764717c0855470dd5680a61a34e6f"} Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.861919 4873 scope.go:117] "RemoveContainer" containerID="31ea34e79d7e4ab84a65b4e2e8527dc55075f1f01fecb83ec4c54c0c1c6f1fc1" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.862718 4873 scope.go:117] "RemoveContainer" containerID="d41610d208e2751d8f43b01b041454558f7764717c0855470dd5680a61a34e6f" Dec 01 08:41:10 crc kubenswrapper[4873]: E1201 08:41:10.862944 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dxlvk_openshift-ovn-kubernetes(a308045b-f54d-42ec-8fdf-5bc3e54ef363)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.884325 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f7ac0b4-d048-46e4-bcee-4a271503e46d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8208cd5fda4df2e3b5a8bd90a18c8aca62b16cc2c60e06f52411f24f3d60dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba3cfd8a611c687414d7ecd6b02b423b351d7ac2a4d041085d4ca3bb2b0709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ae6257c73f3632c499b5f107178f64a9cac2f4325c57b41aa75d6d4ea14dd12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5088b5bf7bc847f59ca9efbf400855a70b99f1bbbf207eb93576c3b3eabae5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:10Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.899826 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:10Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.914415 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36539974b64191f4ca087bacbd02333f90c89fc8273ab7285e9a1d840155707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:10Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.927209 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef7b114-0e07-402d-a37b-315c36011f4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26e4a6e0a8877a213e2e974d4f7a3ad84938ae1365fa45ca7621bc050470e7e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67d090dbd4e672a6d3e777da244480eaf270518c68f8eaf2292ef7c366b8da4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-scwpp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:10Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.936287 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.936340 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.936361 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.936388 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.936408 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:10Z","lastTransitionTime":"2025-12-01T08:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.939412 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfqcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06182c66-4be6-4e00-9a17-4917badfc997\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa8ef5ef07129ec4a7b4cf4738a01cd6444e9bf548134a9da493c7149301387\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfqcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:10Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.956615 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5smrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdde424-6124-432b-9902-cedb9dca26c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6c098fa4f45096e671331702f2995119b74cc8c87c5fa1b175099070bc9099\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lvr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cfdfa53bea0d7e308f7673baf05e104ad7967e0473e36a74a17825433a17c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lvr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5smrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:10Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:10 crc kubenswrapper[4873]: I1201 08:41:10.970699 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9vq2r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08d02e64-a000-4f11-837e-82ad06d925f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzwvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzwvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9vq2r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:10Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.001522 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a2348f-009e-4cc1-b591-e1c14ff0faf2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff98ccdc9925e44e62f0d6e826ea7926dfa2a7d83d33e475a092f128b7fe920a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4676d506268107b87ebe2e21ba6e65d959a23a4d6f38bf0de588bd9c5d15e207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4751109230b67ad8948a2de4794571581d662bfd8abc6dc5c198dbc411172860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3acd19dc33ad40d8857d28bc0ccee4c99a6927e92e5029873234ecb5dc6ed18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7242d1034bd1938cb0bd750cab74db52d454ddd8af75f1e9e0a3c007dcdddd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:10Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.022009 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88068a1cf89e2c507b9d93d305db6af5af196bd9eec805f42d91b6fe090aba35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:11Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.039419 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.039511 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.039536 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.039569 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.039595 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:11Z","lastTransitionTime":"2025-12-01T08:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.041405 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:11Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.060623 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:11Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.090698 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a308045b-f54d-42ec-8fdf-5bc3e54ef363\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d41610d208e2751d8f43b01b041454558f7764717c0855470dd5680a61a34e6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31ea34e79d7e4ab84a65b4e2e8527dc55075f1f01fecb83ec4c54c0c1c6f1fc1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:40:57Z\\\",\\\"message\\\":\\\"ler 9 for removal\\\\nI1201 08:40:57.266189 6351 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 08:40:57.266200 6351 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 08:40:57.266234 6351 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 08:40:57.266249 6351 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1201 08:40:57.266258 6351 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 08:40:57.266274 6351 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 08:40:57.266284 6351 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 08:40:57.266314 6351 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1201 08:40:57.266318 6351 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 08:40:57.266366 6351 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 08:40:57.266388 6351 factory.go:656] Stopping watch factory\\\\nI1201 08:40:57.266393 6351 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 08:40:57.266407 6351 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1201 08:40:57.266399 6351 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 08:40:57.266449 6351 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d41610d208e2751d8f43b01b041454558f7764717c0855470dd5680a61a34e6f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"message\\\":\\\" lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-image-registry/image-registry_TCP_cluster\\\\\\\", UUID:\\\\\\\"83c1e277-3d22-42ae-a355-f7a0ff0bd171\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-image-registry/image-registry\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-image-registry/image-registry_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-image-registry/image-registry\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.93\\\\\\\", Port:5000, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1201 08:41:10.392448 6540 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:41:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dxlvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:11Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.112185 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1c8f149-f188-4759-b2dc-9dc15be1bb13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:40:37.024907 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:40:37.026665 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1609525021/tls.crt::/tmp/serving-cert-1609525021/tls.key\\\\\\\"\\\\nI1201 08:40:42.169070 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:40:42.177896 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:40:42.177940 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:40:42.177968 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:40:42.177975 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:40:42.184767 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1201 08:40:42.184781 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1201 08:40:42.184802 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184808 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184814 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:40:42.184819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:40:42.184822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:40:42.184826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1201 08:40:42.187129 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:11Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.132391 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tpwwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13a2246b-93bb-4586-98ee-53fc84aaae02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dfcc763fe6ab69cab09f09c92ad9a85a4ded49c1443904134f78d1ca0092dae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zfr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tpwwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:11Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.143665 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.143724 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.143734 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.143760 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.143771 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:11Z","lastTransitionTime":"2025-12-01T08:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.154575 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90f08b6fa658ce25fbc25d8d44d956cdb6424a0b047205f19d1a6fd6870b1e90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8d27e3678f1f63f55bd1b76668e328f49368d67d4d7338a6326cb73d13e08f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:11Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.171599 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qpr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c42ab88e96b59d9430a69343e06a329447ae0163cb0f5b9ef2bcafa1bf356583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qpr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:11Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.189295 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8570b876-a13b-4b8f-99c4-4900cfe19a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0afb7d55f395ee39aa245516631b88aab92399973a628f91172c9e410dcc13dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shkn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:11Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.247598 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.247666 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.247686 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.247715 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.247736 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:11Z","lastTransitionTime":"2025-12-01T08:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.351371 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.351448 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.351460 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.351505 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.351517 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:11Z","lastTransitionTime":"2025-12-01T08:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.454868 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.454947 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.454959 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.454984 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.455001 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:11Z","lastTransitionTime":"2025-12-01T08:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.557453 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.557514 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.557527 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.557547 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.557562 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:11Z","lastTransitionTime":"2025-12-01T08:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.661294 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.661360 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.661378 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.661405 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.661426 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:11Z","lastTransitionTime":"2025-12-01T08:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.764233 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.764304 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.764328 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.764356 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.764379 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:11Z","lastTransitionTime":"2025-12-01T08:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.866826 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.866875 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.866886 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.866902 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.866915 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:11Z","lastTransitionTime":"2025-12-01T08:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.868830 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dxlvk_a308045b-f54d-42ec-8fdf-5bc3e54ef363/ovnkube-controller/2.log" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.873303 4873 scope.go:117] "RemoveContainer" containerID="d41610d208e2751d8f43b01b041454558f7764717c0855470dd5680a61a34e6f" Dec 01 08:41:11 crc kubenswrapper[4873]: E1201 08:41:11.873565 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dxlvk_openshift-ovn-kubernetes(a308045b-f54d-42ec-8fdf-5bc3e54ef363)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.891641 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f7ac0b4-d048-46e4-bcee-4a271503e46d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8208cd5fda4df2e3b5a8bd90a18c8aca62b16cc2c60e06f52411f24f3d60dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba3cfd8a611c687414d7ecd6b02b423b351d7ac2a4d041085d4ca3bb2b0709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ae6257c73f3632c499b5f107178f64a9cac2f4325c57b41aa75d6d4ea14dd12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5088b5bf7bc847f59ca9efbf400855a70b99f1bbbf207eb93576c3b3eabae5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:11Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.917501 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:11Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.938751 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36539974b64191f4ca087bacbd02333f90c89fc8273ab7285e9a1d840155707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:11Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.956459 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef7b114-0e07-402d-a37b-315c36011f4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26e4a6e0a8877a213e2e974d4f7a3ad84938ae1365fa45ca7621bc050470e7e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67d090dbd4e672a6d3e777da244480eaf270518c68f8eaf2292ef7c366b8da4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-scwpp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:11Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.971121 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.971174 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.971186 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.971206 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.971219 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:11Z","lastTransitionTime":"2025-12-01T08:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.973355 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfqcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06182c66-4be6-4e00-9a17-4917badfc997\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa8ef5ef07129ec4a7b4cf4738a01cd6444e9bf548134a9da493c7149301387\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfqcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:11Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:11 crc kubenswrapper[4873]: I1201 08:41:11.990835 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5smrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdde424-6124-432b-9902-cedb9dca26c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6c098fa4f45096e671331702f2995119b74cc8c87c5fa1b175099070bc9099\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lvr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cfdfa53bea0d7e308f7673baf05e104ad7967e0473e36a74a17825433a17c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lvr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5smrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:11Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.003951 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9vq2r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08d02e64-a000-4f11-837e-82ad06d925f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzwvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzwvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9vq2r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:12Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.034259 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a2348f-009e-4cc1-b591-e1c14ff0faf2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff98ccdc9925e44e62f0d6e826ea7926dfa2a7d83d33e475a092f128b7fe920a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4676d506268107b87ebe2e21ba6e65d959a23a4d6f38bf0de588bd9c5d15e207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4751109230b67ad8948a2de4794571581d662bfd8abc6dc5c198dbc411172860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3acd19dc33ad40d8857d28bc0ccee4c99a6927e92e5029873234ecb5dc6ed18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7242d1034bd1938cb0bd750cab74db52d454ddd8af75f1e9e0a3c007dcdddd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:12Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.051458 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88068a1cf89e2c507b9d93d305db6af5af196bd9eec805f42d91b6fe090aba35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:12Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.070706 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:12Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.073434 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.073466 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.073476 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.073492 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.073503 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:12Z","lastTransitionTime":"2025-12-01T08:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.084637 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:12Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.112091 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a308045b-f54d-42ec-8fdf-5bc3e54ef363\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d41610d208e2751d8f43b01b041454558f7764717c0855470dd5680a61a34e6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d41610d208e2751d8f43b01b041454558f7764717c0855470dd5680a61a34e6f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"message\\\":\\\" lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-image-registry/image-registry_TCP_cluster\\\\\\\", UUID:\\\\\\\"83c1e277-3d22-42ae-a355-f7a0ff0bd171\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-image-registry/image-registry\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-image-registry/image-registry_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-image-registry/image-registry\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.93\\\\\\\", Port:5000, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1201 08:41:10.392448 6540 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:41:09Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dxlvk_openshift-ovn-kubernetes(a308045b-f54d-42ec-8fdf-5bc3e54ef363)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dxlvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:12Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.125034 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1c8f149-f188-4759-b2dc-9dc15be1bb13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:40:37.024907 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:40:37.026665 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1609525021/tls.crt::/tmp/serving-cert-1609525021/tls.key\\\\\\\"\\\\nI1201 08:40:42.169070 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:40:42.177896 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:40:42.177940 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:40:42.177968 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:40:42.177975 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:40:42.184767 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1201 08:40:42.184781 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1201 08:40:42.184802 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184808 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184814 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:40:42.184819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:40:42.184822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:40:42.184826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1201 08:40:42.187129 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:12Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.139234 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tpwwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13a2246b-93bb-4586-98ee-53fc84aaae02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dfcc763fe6ab69cab09f09c92ad9a85a4ded49c1443904134f78d1ca0092dae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zfr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tpwwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:12Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.151996 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90f08b6fa658ce25fbc25d8d44d956cdb6424a0b047205f19d1a6fd6870b1e90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8d27e3678f1f63f55bd1b76668e328f49368d67d4d7338a6326cb73d13e08f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:12Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.164151 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qpr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c42ab88e96b59d9430a69343e06a329447ae0163cb0f5b9ef2bcafa1bf356583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qpr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:12Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.176459 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.176512 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.176527 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.176543 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.176555 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:12Z","lastTransitionTime":"2025-12-01T08:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.179258 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8570b876-a13b-4b8f-99c4-4900cfe19a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0afb7d55f395ee39aa245516631b88aab92399973a628f91172c9e410dcc13dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shkn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:12Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.278776 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.278851 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.278873 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.278904 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.278927 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:12Z","lastTransitionTime":"2025-12-01T08:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.382338 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.382445 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.382472 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.382503 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.382527 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:12Z","lastTransitionTime":"2025-12-01T08:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.429119 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.429211 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.429246 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:12 crc kubenswrapper[4873]: E1201 08:41:12.429307 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.429350 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:12 crc kubenswrapper[4873]: E1201 08:41:12.429487 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:12 crc kubenswrapper[4873]: E1201 08:41:12.429646 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:12 crc kubenswrapper[4873]: E1201 08:41:12.429712 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.485639 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.485700 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.485722 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.485756 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.485780 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:12Z","lastTransitionTime":"2025-12-01T08:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.590011 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.590110 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.590132 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.590160 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.590180 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:12Z","lastTransitionTime":"2025-12-01T08:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.693183 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.693248 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.693266 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.693290 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.693307 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:12Z","lastTransitionTime":"2025-12-01T08:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.796412 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.796466 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.796481 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.796505 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.796518 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:12Z","lastTransitionTime":"2025-12-01T08:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.899209 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.899274 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.899285 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.899300 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:12 crc kubenswrapper[4873]: I1201 08:41:12.899310 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:12Z","lastTransitionTime":"2025-12-01T08:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.003148 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.003223 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.003243 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.003269 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.003288 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:13Z","lastTransitionTime":"2025-12-01T08:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.106272 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.106356 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.106379 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.106407 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.106430 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:13Z","lastTransitionTime":"2025-12-01T08:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.209614 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.210137 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.210241 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.210336 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.210491 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:13Z","lastTransitionTime":"2025-12-01T08:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.313562 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.313621 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.313638 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.313662 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.313682 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:13Z","lastTransitionTime":"2025-12-01T08:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.416373 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.416488 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.416525 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.416555 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.416576 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:13Z","lastTransitionTime":"2025-12-01T08:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.519751 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.519802 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.519817 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.519836 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.519852 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:13Z","lastTransitionTime":"2025-12-01T08:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.623923 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.623997 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.624049 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.624078 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.624096 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:13Z","lastTransitionTime":"2025-12-01T08:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.727573 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.727661 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.727687 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.727717 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.727742 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:13Z","lastTransitionTime":"2025-12-01T08:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.831570 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.831664 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.831690 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.831866 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.831897 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:13Z","lastTransitionTime":"2025-12-01T08:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.935254 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.935309 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.935325 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.935345 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:13 crc kubenswrapper[4873]: I1201 08:41:13.935360 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:13Z","lastTransitionTime":"2025-12-01T08:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.039069 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.039149 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.039176 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.039206 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.039228 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:14Z","lastTransitionTime":"2025-12-01T08:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.142588 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.142654 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.142677 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.142710 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.142733 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:14Z","lastTransitionTime":"2025-12-01T08:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.247917 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.248772 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.248844 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.248880 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.248904 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:14Z","lastTransitionTime":"2025-12-01T08:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.311787 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.311951 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:14 crc kubenswrapper[4873]: E1201 08:41:14.312008 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:41:46.311976679 +0000 UTC m=+82.214085258 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.312094 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.312178 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.312216 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:14 crc kubenswrapper[4873]: E1201 08:41:14.312183 4873 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:41:14 crc kubenswrapper[4873]: E1201 08:41:14.312348 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:41:46.312322868 +0000 UTC m=+82.214431447 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:41:14 crc kubenswrapper[4873]: E1201 08:41:14.312355 4873 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:41:14 crc kubenswrapper[4873]: E1201 08:41:14.312378 4873 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:41:14 crc kubenswrapper[4873]: E1201 08:41:14.312247 4873 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:41:14 crc kubenswrapper[4873]: E1201 08:41:14.312543 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:41:46.312501853 +0000 UTC m=+82.214610422 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:41:14 crc kubenswrapper[4873]: E1201 08:41:14.312394 4873 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:41:14 crc kubenswrapper[4873]: E1201 08:41:14.312625 4873 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:41:14 crc kubenswrapper[4873]: E1201 08:41:14.312647 4873 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:41:14 crc kubenswrapper[4873]: E1201 08:41:14.312704 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 08:41:46.312687318 +0000 UTC m=+82.214795897 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:41:14 crc kubenswrapper[4873]: E1201 08:41:14.312402 4873 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:41:14 crc kubenswrapper[4873]: E1201 08:41:14.312824 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 08:41:46.3127947 +0000 UTC m=+82.214903279 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.351815 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.351878 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.351896 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.351923 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.351943 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:14Z","lastTransitionTime":"2025-12-01T08:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.429917 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.429953 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.430125 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:41:14 crc kubenswrapper[4873]: E1201 08:41:14.430306 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.430320 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:14 crc kubenswrapper[4873]: E1201 08:41:14.430435 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:14 crc kubenswrapper[4873]: E1201 08:41:14.430548 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:14 crc kubenswrapper[4873]: E1201 08:41:14.430655 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.455896 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.455955 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.455973 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.455999 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.456048 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:14Z","lastTransitionTime":"2025-12-01T08:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.457202 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1c8f149-f188-4759-b2dc-9dc15be1bb13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:40:37.024907 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:40:37.026665 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1609525021/tls.crt::/tmp/serving-cert-1609525021/tls.key\\\\\\\"\\\\nI1201 08:40:42.169070 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:40:42.177896 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:40:42.177940 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:40:42.177968 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:40:42.177975 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:40:42.184767 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1201 08:40:42.184781 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1201 08:40:42.184802 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184808 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184814 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:40:42.184819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:40:42.184822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:40:42.184826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1201 08:40:42.187129 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.480187 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tpwwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13a2246b-93bb-4586-98ee-53fc84aaae02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dfcc763fe6ab69cab09f09c92ad9a85a4ded49c1443904134f78d1ca0092dae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zfr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tpwwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.501884 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90f08b6fa658ce25fbc25d8d44d956cdb6424a0b047205f19d1a6fd6870b1e90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8d27e3678f1f63f55bd1b76668e328f49368d67d4d7338a6326cb73d13e08f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.519630 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qpr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c42ab88e96b59d9430a69343e06a329447ae0163cb0f5b9ef2bcafa1bf356583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qpr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.548867 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8570b876-a13b-4b8f-99c4-4900cfe19a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0afb7d55f395ee39aa245516631b88aab92399973a628f91172c9e410dcc13dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shkn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.558810 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.558852 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.558865 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.558890 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.558904 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:14Z","lastTransitionTime":"2025-12-01T08:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.594461 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfqcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06182c66-4be6-4e00-9a17-4917badfc997\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa8ef5ef07129ec4a7b4cf4738a01cd6444e9bf548134a9da493c7149301387\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfqcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.615074 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/08d02e64-a000-4f11-837e-82ad06d925f5-metrics-certs\") pod \"network-metrics-daemon-9vq2r\" (UID: \"08d02e64-a000-4f11-837e-82ad06d925f5\") " pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:41:14 crc kubenswrapper[4873]: E1201 08:41:14.615234 4873 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 08:41:14 crc kubenswrapper[4873]: E1201 08:41:14.615327 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/08d02e64-a000-4f11-837e-82ad06d925f5-metrics-certs podName:08d02e64-a000-4f11-837e-82ad06d925f5 nodeName:}" failed. No retries permitted until 2025-12-01 08:41:30.615303087 +0000 UTC m=+66.517411626 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/08d02e64-a000-4f11-837e-82ad06d925f5-metrics-certs") pod "network-metrics-daemon-9vq2r" (UID: "08d02e64-a000-4f11-837e-82ad06d925f5") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.623482 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f7ac0b4-d048-46e4-bcee-4a271503e46d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8208cd5fda4df2e3b5a8bd90a18c8aca62b16cc2c60e06f52411f24f3d60dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba3cfd8a611c687414d7ecd6b02b423b351d7ac2a4d041085d4ca3bb2b0709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ae6257c73f3632c499b5f107178f64a9cac2f4325c57b41aa75d6d4ea14dd12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5088b5bf7bc847f59ca9efbf400855a70b99f1bbbf207eb93576c3b3eabae5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.635568 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.647935 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36539974b64191f4ca087bacbd02333f90c89fc8273ab7285e9a1d840155707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.662158 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.662201 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.662210 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.662227 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.662239 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:14Z","lastTransitionTime":"2025-12-01T08:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.662784 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef7b114-0e07-402d-a37b-315c36011f4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26e4a6e0a8877a213e2e974d4f7a3ad84938ae1365fa45ca7621bc050470e7e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67d090dbd4e672a6d3e777da244480eaf270518c68f8eaf2292ef7c366b8da4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-scwpp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.691518 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a308045b-f54d-42ec-8fdf-5bc3e54ef363\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d41610d208e2751d8f43b01b041454558f7764717c0855470dd5680a61a34e6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d41610d208e2751d8f43b01b041454558f7764717c0855470dd5680a61a34e6f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"message\\\":\\\" lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-image-registry/image-registry_TCP_cluster\\\\\\\", UUID:\\\\\\\"83c1e277-3d22-42ae-a355-f7a0ff0bd171\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-image-registry/image-registry\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-image-registry/image-registry_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-image-registry/image-registry\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.93\\\\\\\", Port:5000, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1201 08:41:10.392448 6540 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:41:09Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dxlvk_openshift-ovn-kubernetes(a308045b-f54d-42ec-8fdf-5bc3e54ef363)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dxlvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.710158 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5smrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdde424-6124-432b-9902-cedb9dca26c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6c098fa4f45096e671331702f2995119b74cc8c87c5fa1b175099070bc9099\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lvr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cfdfa53bea0d7e308f7673baf05e104ad7967e0473e36a74a17825433a17c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lvr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5smrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.726961 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9vq2r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08d02e64-a000-4f11-837e-82ad06d925f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzwvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzwvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9vq2r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.762930 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a2348f-009e-4cc1-b591-e1c14ff0faf2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff98ccdc9925e44e62f0d6e826ea7926dfa2a7d83d33e475a092f128b7fe920a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4676d506268107b87ebe2e21ba6e65d959a23a4d6f38bf0de588bd9c5d15e207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4751109230b67ad8948a2de4794571581d662bfd8abc6dc5c198dbc411172860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3acd19dc33ad40d8857d28bc0ccee4c99a6927e92e5029873234ecb5dc6ed18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7242d1034bd1938cb0bd750cab74db52d454ddd8af75f1e9e0a3c007dcdddd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.764297 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.764372 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.764393 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.764422 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.764446 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:14Z","lastTransitionTime":"2025-12-01T08:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.785898 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88068a1cf89e2c507b9d93d305db6af5af196bd9eec805f42d91b6fe090aba35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.805089 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.826110 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.868863 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.868935 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.868949 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.868972 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.868987 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:14Z","lastTransitionTime":"2025-12-01T08:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.972550 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.972639 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.972653 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.972679 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:14 crc kubenswrapper[4873]: I1201 08:41:14.972695 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:14Z","lastTransitionTime":"2025-12-01T08:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:15 crc kubenswrapper[4873]: I1201 08:41:15.076905 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:15 crc kubenswrapper[4873]: I1201 08:41:15.076999 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:15 crc kubenswrapper[4873]: I1201 08:41:15.077073 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:15 crc kubenswrapper[4873]: I1201 08:41:15.077111 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:15 crc kubenswrapper[4873]: I1201 08:41:15.077136 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:15Z","lastTransitionTime":"2025-12-01T08:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:15 crc kubenswrapper[4873]: I1201 08:41:15.181711 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:15 crc kubenswrapper[4873]: I1201 08:41:15.181781 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:15 crc kubenswrapper[4873]: I1201 08:41:15.181805 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:15 crc kubenswrapper[4873]: I1201 08:41:15.181833 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:15 crc kubenswrapper[4873]: I1201 08:41:15.181854 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:15Z","lastTransitionTime":"2025-12-01T08:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:15 crc kubenswrapper[4873]: I1201 08:41:15.285745 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:15 crc kubenswrapper[4873]: I1201 08:41:15.285811 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:15 crc kubenswrapper[4873]: I1201 08:41:15.285829 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:15 crc kubenswrapper[4873]: I1201 08:41:15.285861 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:15 crc kubenswrapper[4873]: I1201 08:41:15.285880 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:15Z","lastTransitionTime":"2025-12-01T08:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:15 crc kubenswrapper[4873]: I1201 08:41:15.389455 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:15 crc kubenswrapper[4873]: I1201 08:41:15.389563 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:15 crc kubenswrapper[4873]: I1201 08:41:15.389605 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:15 crc kubenswrapper[4873]: I1201 08:41:15.389638 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:15 crc kubenswrapper[4873]: I1201 08:41:15.389657 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:15Z","lastTransitionTime":"2025-12-01T08:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:15 crc kubenswrapper[4873]: I1201 08:41:15.492760 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:15 crc kubenswrapper[4873]: I1201 08:41:15.492818 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:15 crc kubenswrapper[4873]: I1201 08:41:15.492836 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:15 crc kubenswrapper[4873]: I1201 08:41:15.492859 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:15 crc kubenswrapper[4873]: I1201 08:41:15.492876 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:15Z","lastTransitionTime":"2025-12-01T08:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:15 crc kubenswrapper[4873]: I1201 08:41:15.595555 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:15 crc kubenswrapper[4873]: I1201 08:41:15.595631 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:15 crc kubenswrapper[4873]: I1201 08:41:15.595645 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:15 crc kubenswrapper[4873]: I1201 08:41:15.595666 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:15 crc kubenswrapper[4873]: I1201 08:41:15.595680 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:15Z","lastTransitionTime":"2025-12-01T08:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:15 crc kubenswrapper[4873]: I1201 08:41:15.699433 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:15 crc kubenswrapper[4873]: I1201 08:41:15.699478 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:15 crc kubenswrapper[4873]: I1201 08:41:15.699488 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:15 crc kubenswrapper[4873]: I1201 08:41:15.699505 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:15 crc kubenswrapper[4873]: I1201 08:41:15.699517 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:15Z","lastTransitionTime":"2025-12-01T08:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:15 crc kubenswrapper[4873]: I1201 08:41:15.802833 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:15 crc kubenswrapper[4873]: I1201 08:41:15.802880 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:15 crc kubenswrapper[4873]: I1201 08:41:15.802906 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:15 crc kubenswrapper[4873]: I1201 08:41:15.802932 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:15 crc kubenswrapper[4873]: I1201 08:41:15.802948 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:15Z","lastTransitionTime":"2025-12-01T08:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:15 crc kubenswrapper[4873]: I1201 08:41:15.906305 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:15 crc kubenswrapper[4873]: I1201 08:41:15.906380 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:15 crc kubenswrapper[4873]: I1201 08:41:15.906396 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:15 crc kubenswrapper[4873]: I1201 08:41:15.906418 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:15 crc kubenswrapper[4873]: I1201 08:41:15.906443 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:15Z","lastTransitionTime":"2025-12-01T08:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.002805 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.008995 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.009049 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.009060 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.009076 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.009088 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:16Z","lastTransitionTime":"2025-12-01T08:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.017195 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.022397 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1c8f149-f188-4759-b2dc-9dc15be1bb13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:40:37.024907 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:40:37.026665 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1609525021/tls.crt::/tmp/serving-cert-1609525021/tls.key\\\\\\\"\\\\nI1201 08:40:42.169070 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:40:42.177896 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:40:42.177940 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:40:42.177968 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:40:42.177975 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:40:42.184767 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1201 08:40:42.184781 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1201 08:40:42.184802 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184808 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184814 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:40:42.184819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:40:42.184822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:40:42.184826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1201 08:40:42.187129 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.041859 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tpwwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13a2246b-93bb-4586-98ee-53fc84aaae02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dfcc763fe6ab69cab09f09c92ad9a85a4ded49c1443904134f78d1ca0092dae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zfr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tpwwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.058314 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90f08b6fa658ce25fbc25d8d44d956cdb6424a0b047205f19d1a6fd6870b1e90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8d27e3678f1f63f55bd1b76668e328f49368d67d4d7338a6326cb73d13e08f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.072188 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qpr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c42ab88e96b59d9430a69343e06a329447ae0163cb0f5b9ef2bcafa1bf356583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qpr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.093315 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8570b876-a13b-4b8f-99c4-4900cfe19a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0afb7d55f395ee39aa245516631b88aab92399973a628f91172c9e410dcc13dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shkn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.107187 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f7ac0b4-d048-46e4-bcee-4a271503e46d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8208cd5fda4df2e3b5a8bd90a18c8aca62b16cc2c60e06f52411f24f3d60dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba3cfd8a611c687414d7ecd6b02b423b351d7ac2a4d041085d4ca3bb2b0709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ae6257c73f3632c499b5f107178f64a9cac2f4325c57b41aa75d6d4ea14dd12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5088b5bf7bc847f59ca9efbf400855a70b99f1bbbf207eb93576c3b3eabae5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.113090 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.113144 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.113163 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.113188 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.113209 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:16Z","lastTransitionTime":"2025-12-01T08:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.121721 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.140593 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36539974b64191f4ca087bacbd02333f90c89fc8273ab7285e9a1d840155707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.155000 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef7b114-0e07-402d-a37b-315c36011f4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26e4a6e0a8877a213e2e974d4f7a3ad84938ae1365fa45ca7621bc050470e7e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67d090dbd4e672a6d3e777da244480eaf270518c68f8eaf2292ef7c366b8da4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-scwpp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.165570 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfqcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06182c66-4be6-4e00-9a17-4917badfc997\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa8ef5ef07129ec4a7b4cf4738a01cd6444e9bf548134a9da493c7149301387\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfqcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.179745 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5smrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdde424-6124-432b-9902-cedb9dca26c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6c098fa4f45096e671331702f2995119b74cc8c87c5fa1b175099070bc9099\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lvr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cfdfa53bea0d7e308f7673baf05e104ad7967e0473e36a74a17825433a17c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lvr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5smrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.195826 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9vq2r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08d02e64-a000-4f11-837e-82ad06d925f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzwvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzwvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9vq2r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.211705 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.211778 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.211791 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.211816 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.211831 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:16Z","lastTransitionTime":"2025-12-01T08:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:16 crc kubenswrapper[4873]: E1201 08:41:16.232333 4873 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963d6124-09f6-48b8-b38c-854877aa92e3\\\",\\\"systemUUID\\\":\\\"27c66b0b-2450-400c-b6ea-cdb9cbbf95ad\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.233543 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a2348f-009e-4cc1-b591-e1c14ff0faf2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff98ccdc9925e44e62f0d6e826ea7926dfa2a7d83d33e475a092f128b7fe920a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4676d506268107b87ebe2e21ba6e65d959a23a4d6f38bf0de588bd9c5d15e207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4751109230b67ad8948a2de4794571581d662bfd8abc6dc5c198dbc411172860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3acd19dc33ad40d8857d28bc0ccee4c99a6927e92e5029873234ecb5dc6ed18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7242d1034bd1938cb0bd750cab74db52d454ddd8af75f1e9e0a3c007dcdddd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.238098 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.238175 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.238195 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.238222 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.238243 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:16Z","lastTransitionTime":"2025-12-01T08:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.256472 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88068a1cf89e2c507b9d93d305db6af5af196bd9eec805f42d91b6fe090aba35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:16 crc kubenswrapper[4873]: E1201 08:41:16.260413 4873 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963d6124-09f6-48b8-b38c-854877aa92e3\\\",\\\"systemUUID\\\":\\\"27c66b0b-2450-400c-b6ea-cdb9cbbf95ad\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.266145 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.266282 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.266308 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.266338 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.266388 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:16Z","lastTransitionTime":"2025-12-01T08:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.278228 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:16 crc kubenswrapper[4873]: E1201 08:41:16.290606 4873 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963d6124-09f6-48b8-b38c-854877aa92e3\\\",\\\"systemUUID\\\":\\\"27c66b0b-2450-400c-b6ea-cdb9cbbf95ad\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.296345 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.296445 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.296460 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.296485 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.296499 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:16Z","lastTransitionTime":"2025-12-01T08:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.300967 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:16 crc kubenswrapper[4873]: E1201 08:41:16.316777 4873 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963d6124-09f6-48b8-b38c-854877aa92e3\\\",\\\"systemUUID\\\":\\\"27c66b0b-2450-400c-b6ea-cdb9cbbf95ad\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.322936 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.323049 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.323089 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.323122 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.323143 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:16Z","lastTransitionTime":"2025-12-01T08:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.335498 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a308045b-f54d-42ec-8fdf-5bc3e54ef363\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d41610d208e2751d8f43b01b041454558f7764717c0855470dd5680a61a34e6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d41610d208e2751d8f43b01b041454558f7764717c0855470dd5680a61a34e6f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"message\\\":\\\" lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-image-registry/image-registry_TCP_cluster\\\\\\\", UUID:\\\\\\\"83c1e277-3d22-42ae-a355-f7a0ff0bd171\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-image-registry/image-registry\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-image-registry/image-registry_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-image-registry/image-registry\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.93\\\\\\\", Port:5000, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1201 08:41:10.392448 6540 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:41:09Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dxlvk_openshift-ovn-kubernetes(a308045b-f54d-42ec-8fdf-5bc3e54ef363)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dxlvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:16 crc kubenswrapper[4873]: E1201 08:41:16.345145 4873 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963d6124-09f6-48b8-b38c-854877aa92e3\\\",\\\"systemUUID\\\":\\\"27c66b0b-2450-400c-b6ea-cdb9cbbf95ad\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:16 crc kubenswrapper[4873]: E1201 08:41:16.345396 4873 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.347734 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.347790 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.347812 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.347839 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.347858 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:16Z","lastTransitionTime":"2025-12-01T08:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.429236 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.429297 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.429368 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.429521 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:16 crc kubenswrapper[4873]: E1201 08:41:16.429510 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:16 crc kubenswrapper[4873]: E1201 08:41:16.429666 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:41:16 crc kubenswrapper[4873]: E1201 08:41:16.429959 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:16 crc kubenswrapper[4873]: E1201 08:41:16.430096 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.450935 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.451057 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.451086 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.451117 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.451143 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:16Z","lastTransitionTime":"2025-12-01T08:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.554492 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.554576 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.554599 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.554631 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.554653 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:16Z","lastTransitionTime":"2025-12-01T08:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.664150 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.664265 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.665001 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.665107 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.665133 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:16Z","lastTransitionTime":"2025-12-01T08:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.768584 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.768661 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.768684 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.768716 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.768739 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:16Z","lastTransitionTime":"2025-12-01T08:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.872448 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.872518 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.872531 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.872551 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.872565 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:16Z","lastTransitionTime":"2025-12-01T08:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.976067 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.976125 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.976139 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.976160 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:16 crc kubenswrapper[4873]: I1201 08:41:16.976174 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:16Z","lastTransitionTime":"2025-12-01T08:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:17 crc kubenswrapper[4873]: I1201 08:41:17.079227 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:17 crc kubenswrapper[4873]: I1201 08:41:17.079301 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:17 crc kubenswrapper[4873]: I1201 08:41:17.079325 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:17 crc kubenswrapper[4873]: I1201 08:41:17.079357 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:17 crc kubenswrapper[4873]: I1201 08:41:17.079376 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:17Z","lastTransitionTime":"2025-12-01T08:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:17 crc kubenswrapper[4873]: I1201 08:41:17.183188 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:17 crc kubenswrapper[4873]: I1201 08:41:17.183269 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:17 crc kubenswrapper[4873]: I1201 08:41:17.183291 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:17 crc kubenswrapper[4873]: I1201 08:41:17.183325 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:17 crc kubenswrapper[4873]: I1201 08:41:17.183348 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:17Z","lastTransitionTime":"2025-12-01T08:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:17 crc kubenswrapper[4873]: I1201 08:41:17.286089 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:17 crc kubenswrapper[4873]: I1201 08:41:17.286176 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:17 crc kubenswrapper[4873]: I1201 08:41:17.286194 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:17 crc kubenswrapper[4873]: I1201 08:41:17.286220 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:17 crc kubenswrapper[4873]: I1201 08:41:17.286240 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:17Z","lastTransitionTime":"2025-12-01T08:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:17 crc kubenswrapper[4873]: I1201 08:41:17.388749 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:17 crc kubenswrapper[4873]: I1201 08:41:17.388815 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:17 crc kubenswrapper[4873]: I1201 08:41:17.388839 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:17 crc kubenswrapper[4873]: I1201 08:41:17.388868 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:17 crc kubenswrapper[4873]: I1201 08:41:17.388889 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:17Z","lastTransitionTime":"2025-12-01T08:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:17 crc kubenswrapper[4873]: I1201 08:41:17.491736 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:17 crc kubenswrapper[4873]: I1201 08:41:17.491854 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:17 crc kubenswrapper[4873]: I1201 08:41:17.491878 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:17 crc kubenswrapper[4873]: I1201 08:41:17.491908 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:17 crc kubenswrapper[4873]: I1201 08:41:17.491925 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:17Z","lastTransitionTime":"2025-12-01T08:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:17 crc kubenswrapper[4873]: I1201 08:41:17.595985 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:17 crc kubenswrapper[4873]: I1201 08:41:17.596125 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:17 crc kubenswrapper[4873]: I1201 08:41:17.596146 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:17 crc kubenswrapper[4873]: I1201 08:41:17.596173 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:17 crc kubenswrapper[4873]: I1201 08:41:17.596191 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:17Z","lastTransitionTime":"2025-12-01T08:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:17 crc kubenswrapper[4873]: I1201 08:41:17.699992 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:17 crc kubenswrapper[4873]: I1201 08:41:17.700078 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:17 crc kubenswrapper[4873]: I1201 08:41:17.700097 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:17 crc kubenswrapper[4873]: I1201 08:41:17.700127 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:17 crc kubenswrapper[4873]: I1201 08:41:17.700148 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:17Z","lastTransitionTime":"2025-12-01T08:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:17 crc kubenswrapper[4873]: I1201 08:41:17.804229 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:17 crc kubenswrapper[4873]: I1201 08:41:17.804331 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:17 crc kubenswrapper[4873]: I1201 08:41:17.804358 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:17 crc kubenswrapper[4873]: I1201 08:41:17.804396 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:17 crc kubenswrapper[4873]: I1201 08:41:17.804422 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:17Z","lastTransitionTime":"2025-12-01T08:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:17 crc kubenswrapper[4873]: I1201 08:41:17.908144 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:17 crc kubenswrapper[4873]: I1201 08:41:17.908224 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:17 crc kubenswrapper[4873]: I1201 08:41:17.908249 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:17 crc kubenswrapper[4873]: I1201 08:41:17.908279 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:17 crc kubenswrapper[4873]: I1201 08:41:17.908297 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:17Z","lastTransitionTime":"2025-12-01T08:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.012321 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.012383 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.012404 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.012428 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.012446 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:18Z","lastTransitionTime":"2025-12-01T08:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.116223 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.116310 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.116345 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.116377 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.116399 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:18Z","lastTransitionTime":"2025-12-01T08:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.220788 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.220897 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.220926 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.220960 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.220987 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:18Z","lastTransitionTime":"2025-12-01T08:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.325714 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.325793 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.325812 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.325845 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.325868 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:18Z","lastTransitionTime":"2025-12-01T08:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.429393 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.429451 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.429472 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.429428 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:18 crc kubenswrapper[4873]: E1201 08:41:18.429712 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:18 crc kubenswrapper[4873]: E1201 08:41:18.430045 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:18 crc kubenswrapper[4873]: E1201 08:41:18.430138 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:41:18 crc kubenswrapper[4873]: E1201 08:41:18.430138 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.430437 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.430528 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.430545 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.430569 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.430591 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:18Z","lastTransitionTime":"2025-12-01T08:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.534285 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.534400 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.534410 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.534428 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.534438 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:18Z","lastTransitionTime":"2025-12-01T08:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.639589 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.639650 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.639667 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.639693 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.639711 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:18Z","lastTransitionTime":"2025-12-01T08:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.748806 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.748880 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.748896 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.748931 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.748951 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:18Z","lastTransitionTime":"2025-12-01T08:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.852549 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.852613 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.852630 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.852655 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.852673 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:18Z","lastTransitionTime":"2025-12-01T08:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.955504 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.955594 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.955615 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.955643 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:18 crc kubenswrapper[4873]: I1201 08:41:18.955670 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:18Z","lastTransitionTime":"2025-12-01T08:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.059264 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.059342 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.059361 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.059393 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.059412 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:19Z","lastTransitionTime":"2025-12-01T08:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.162385 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.162436 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.162452 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.162477 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.162496 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:19Z","lastTransitionTime":"2025-12-01T08:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.265979 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.266068 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.266085 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.266109 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.266128 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:19Z","lastTransitionTime":"2025-12-01T08:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.369885 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.369964 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.369989 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.370049 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.370075 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:19Z","lastTransitionTime":"2025-12-01T08:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.474118 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.474176 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.474193 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.474215 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.474233 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:19Z","lastTransitionTime":"2025-12-01T08:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.577873 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.577935 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.577946 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.577968 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.577980 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:19Z","lastTransitionTime":"2025-12-01T08:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.681435 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.681569 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.681589 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.681732 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.681750 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:19Z","lastTransitionTime":"2025-12-01T08:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.784620 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.784692 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.784711 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.784738 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.784757 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:19Z","lastTransitionTime":"2025-12-01T08:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.887730 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.887795 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.887816 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.887844 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.887868 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:19Z","lastTransitionTime":"2025-12-01T08:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.991765 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.991882 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.991902 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.991939 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:19 crc kubenswrapper[4873]: I1201 08:41:19.991958 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:19Z","lastTransitionTime":"2025-12-01T08:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.095644 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.095888 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.095907 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.095931 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.095946 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:20Z","lastTransitionTime":"2025-12-01T08:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.199456 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.199526 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.199544 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.199570 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.199588 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:20Z","lastTransitionTime":"2025-12-01T08:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.303178 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.303294 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.303332 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.303370 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.303396 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:20Z","lastTransitionTime":"2025-12-01T08:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.410811 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.410881 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.410899 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.410925 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.410944 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:20Z","lastTransitionTime":"2025-12-01T08:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.429482 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.429632 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:20 crc kubenswrapper[4873]: E1201 08:41:20.429678 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.429702 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:20 crc kubenswrapper[4873]: E1201 08:41:20.429968 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.429742 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:41:20 crc kubenswrapper[4873]: E1201 08:41:20.430131 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:20 crc kubenswrapper[4873]: E1201 08:41:20.430314 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.514388 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.514492 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.514518 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.514553 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.514579 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:20Z","lastTransitionTime":"2025-12-01T08:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.617473 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.617536 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.617554 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.617577 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.617595 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:20Z","lastTransitionTime":"2025-12-01T08:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.720691 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.720759 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.720782 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.720811 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.720832 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:20Z","lastTransitionTime":"2025-12-01T08:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.824585 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.824651 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.824674 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.824704 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.824725 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:20Z","lastTransitionTime":"2025-12-01T08:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.928007 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.928129 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.928148 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.928183 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:20 crc kubenswrapper[4873]: I1201 08:41:20.928206 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:20Z","lastTransitionTime":"2025-12-01T08:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.031720 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.031797 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.031815 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.031841 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.031860 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:21Z","lastTransitionTime":"2025-12-01T08:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.135971 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.136071 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.136090 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.136124 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.136144 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:21Z","lastTransitionTime":"2025-12-01T08:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.240059 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.240143 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.240166 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.240207 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.240247 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:21Z","lastTransitionTime":"2025-12-01T08:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.345289 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.345364 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.345386 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.345417 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.345443 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:21Z","lastTransitionTime":"2025-12-01T08:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.448951 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.449066 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.449092 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.449120 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.449142 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:21Z","lastTransitionTime":"2025-12-01T08:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.552280 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.552366 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.552385 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.552411 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.552433 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:21Z","lastTransitionTime":"2025-12-01T08:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.680629 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.680717 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.680750 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.680779 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.680800 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:21Z","lastTransitionTime":"2025-12-01T08:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.784492 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.784562 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.784583 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.784608 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.784635 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:21Z","lastTransitionTime":"2025-12-01T08:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.888168 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.888223 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.888241 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.888263 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.888282 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:21Z","lastTransitionTime":"2025-12-01T08:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.991307 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.991363 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.991379 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.991396 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:21 crc kubenswrapper[4873]: I1201 08:41:21.991409 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:21Z","lastTransitionTime":"2025-12-01T08:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.094451 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.094509 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.094522 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.094539 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.094557 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:22Z","lastTransitionTime":"2025-12-01T08:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.198122 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.198186 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.198204 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.198233 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.198258 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:22Z","lastTransitionTime":"2025-12-01T08:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.301054 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.301147 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.301182 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.301217 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.301243 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:22Z","lastTransitionTime":"2025-12-01T08:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.405101 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.405201 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.405227 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.405257 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.405279 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:22Z","lastTransitionTime":"2025-12-01T08:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.430138 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.430181 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.430188 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:41:22 crc kubenswrapper[4873]: E1201 08:41:22.430327 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.430382 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:22 crc kubenswrapper[4873]: E1201 08:41:22.430477 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:22 crc kubenswrapper[4873]: E1201 08:41:22.430569 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:22 crc kubenswrapper[4873]: E1201 08:41:22.431247 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.431742 4873 scope.go:117] "RemoveContainer" containerID="d41610d208e2751d8f43b01b041454558f7764717c0855470dd5680a61a34e6f" Dec 01 08:41:22 crc kubenswrapper[4873]: E1201 08:41:22.432174 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dxlvk_openshift-ovn-kubernetes(a308045b-f54d-42ec-8fdf-5bc3e54ef363)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.508096 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.508174 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.508189 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.508211 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.508225 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:22Z","lastTransitionTime":"2025-12-01T08:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.611712 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.611780 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.611816 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.611850 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.611872 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:22Z","lastTransitionTime":"2025-12-01T08:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.714502 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.714585 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.714604 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.714633 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.714653 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:22Z","lastTransitionTime":"2025-12-01T08:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.817525 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.817613 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.817638 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.817667 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.817692 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:22Z","lastTransitionTime":"2025-12-01T08:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.955978 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.956074 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.956089 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.956107 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:22 crc kubenswrapper[4873]: I1201 08:41:22.956121 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:22Z","lastTransitionTime":"2025-12-01T08:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.058871 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.058929 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.058941 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.058958 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.058970 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:23Z","lastTransitionTime":"2025-12-01T08:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.162354 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.162410 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.162422 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.162438 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.162448 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:23Z","lastTransitionTime":"2025-12-01T08:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.266100 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.266188 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.266203 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.266223 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.266238 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:23Z","lastTransitionTime":"2025-12-01T08:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.370209 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.370275 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.370289 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.370309 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.370324 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:23Z","lastTransitionTime":"2025-12-01T08:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.473652 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.473727 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.473748 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.473775 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.473798 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:23Z","lastTransitionTime":"2025-12-01T08:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.577315 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.577368 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.577383 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.577401 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.577415 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:23Z","lastTransitionTime":"2025-12-01T08:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.680373 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.680420 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.680429 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.680446 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.680457 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:23Z","lastTransitionTime":"2025-12-01T08:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.784335 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.784399 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.784419 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.784446 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.784467 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:23Z","lastTransitionTime":"2025-12-01T08:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.887800 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.887928 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.887949 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.887978 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.887997 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:23Z","lastTransitionTime":"2025-12-01T08:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.991925 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.992010 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.992068 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.992100 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:23 crc kubenswrapper[4873]: I1201 08:41:23.992159 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:23Z","lastTransitionTime":"2025-12-01T08:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.096469 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.096566 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.096583 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.096609 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.096632 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:24Z","lastTransitionTime":"2025-12-01T08:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.211055 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.211140 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.211159 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.211188 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.211207 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:24Z","lastTransitionTime":"2025-12-01T08:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.314874 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.314946 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.314965 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.314990 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.315008 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:24Z","lastTransitionTime":"2025-12-01T08:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.418640 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.418695 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.418712 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.418731 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.418744 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:24Z","lastTransitionTime":"2025-12-01T08:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.429868 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:24 crc kubenswrapper[4873]: E1201 08:41:24.430112 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.430354 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.430391 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.430513 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:24 crc kubenswrapper[4873]: E1201 08:41:24.430603 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:24 crc kubenswrapper[4873]: E1201 08:41:24.430741 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:41:24 crc kubenswrapper[4873]: E1201 08:41:24.430888 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.462801 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a308045b-f54d-42ec-8fdf-5bc3e54ef363\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d41610d208e2751d8f43b01b041454558f7764717c0855470dd5680a61a34e6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d41610d208e2751d8f43b01b041454558f7764717c0855470dd5680a61a34e6f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"message\\\":\\\" lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-image-registry/image-registry_TCP_cluster\\\\\\\", UUID:\\\\\\\"83c1e277-3d22-42ae-a355-f7a0ff0bd171\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-image-registry/image-registry\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-image-registry/image-registry_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-image-registry/image-registry\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.93\\\\\\\", Port:5000, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1201 08:41:10.392448 6540 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:41:09Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dxlvk_openshift-ovn-kubernetes(a308045b-f54d-42ec-8fdf-5bc3e54ef363)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dxlvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:24Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.483501 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5smrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdde424-6124-432b-9902-cedb9dca26c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6c098fa4f45096e671331702f2995119b74cc8c87c5fa1b175099070bc9099\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lvr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cfdfa53bea0d7e308f7673baf05e104ad7967e0473e36a74a17825433a17c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lvr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5smrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:24Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.505497 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9vq2r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08d02e64-a000-4f11-837e-82ad06d925f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzwvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzwvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9vq2r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:24Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.522235 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.522317 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.522337 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.522364 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.522386 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:24Z","lastTransitionTime":"2025-12-01T08:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.545996 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a2348f-009e-4cc1-b591-e1c14ff0faf2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff98ccdc9925e44e62f0d6e826ea7926dfa2a7d83d33e475a092f128b7fe920a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4676d506268107b87ebe2e21ba6e65d959a23a4d6f38bf0de588bd9c5d15e207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4751109230b67ad8948a2de4794571581d662bfd8abc6dc5c198dbc411172860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3acd19dc33ad40d8857d28bc0ccee4c99a6927e92e5029873234ecb5dc6ed18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7242d1034bd1938cb0bd750cab74db52d454ddd8af75f1e9e0a3c007dcdddd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:24Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.568457 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88068a1cf89e2c507b9d93d305db6af5af196bd9eec805f42d91b6fe090aba35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:24Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.591083 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:24Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.614766 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:24Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.629948 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.630001 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.630047 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.630074 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.630098 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:24Z","lastTransitionTime":"2025-12-01T08:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.637874 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6938caf5-0969-4f3a-ba82-ae50109d7be0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25326d274f24d0da05778ed8c5a80d465065df87cf625be2c35d9c95763d4aab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9ffbc1a66dd17bac0e36d480467c72521d7a97c5376dafe57fa15efe571c78b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ce4df67c597a961222280dff543040c444f4d7eace37c5cfe07c7f9d89ea70a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aae94f574fbe3c1c7fcd3dbe889f7aa484e1482cee011ebfb2044151d4ca8eac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aae94f574fbe3c1c7fcd3dbe889f7aa484e1482cee011ebfb2044151d4ca8eac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:24Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.669956 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1c8f149-f188-4759-b2dc-9dc15be1bb13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:40:37.024907 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:40:37.026665 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1609525021/tls.crt::/tmp/serving-cert-1609525021/tls.key\\\\\\\"\\\\nI1201 08:40:42.169070 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:40:42.177896 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:40:42.177940 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:40:42.177968 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:40:42.177975 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:40:42.184767 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1201 08:40:42.184781 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1201 08:40:42.184802 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184808 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184814 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:40:42.184819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:40:42.184822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:40:42.184826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1201 08:40:42.187129 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:24Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.696356 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tpwwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13a2246b-93bb-4586-98ee-53fc84aaae02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dfcc763fe6ab69cab09f09c92ad9a85a4ded49c1443904134f78d1ca0092dae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zfr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tpwwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:24Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.721942 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90f08b6fa658ce25fbc25d8d44d956cdb6424a0b047205f19d1a6fd6870b1e90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8d27e3678f1f63f55bd1b76668e328f49368d67d4d7338a6326cb73d13e08f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:24Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.733318 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.733559 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.733960 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.734385 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.734747 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:24Z","lastTransitionTime":"2025-12-01T08:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.741349 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qpr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c42ab88e96b59d9430a69343e06a329447ae0163cb0f5b9ef2bcafa1bf356583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qpr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:24Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.766913 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8570b876-a13b-4b8f-99c4-4900cfe19a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0afb7d55f395ee39aa245516631b88aab92399973a628f91172c9e410dcc13dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shkn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:24Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.787105 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfqcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06182c66-4be6-4e00-9a17-4917badfc997\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa8ef5ef07129ec4a7b4cf4738a01cd6444e9bf548134a9da493c7149301387\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfqcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:24Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.810993 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f7ac0b4-d048-46e4-bcee-4a271503e46d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8208cd5fda4df2e3b5a8bd90a18c8aca62b16cc2c60e06f52411f24f3d60dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba3cfd8a611c687414d7ecd6b02b423b351d7ac2a4d041085d4ca3bb2b0709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ae6257c73f3632c499b5f107178f64a9cac2f4325c57b41aa75d6d4ea14dd12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5088b5bf7bc847f59ca9efbf400855a70b99f1bbbf207eb93576c3b3eabae5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:24Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.835446 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:24Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.838639 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.838719 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.838743 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.838776 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.838807 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:24Z","lastTransitionTime":"2025-12-01T08:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.857750 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36539974b64191f4ca087bacbd02333f90c89fc8273ab7285e9a1d840155707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:24Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.874631 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef7b114-0e07-402d-a37b-315c36011f4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26e4a6e0a8877a213e2e974d4f7a3ad84938ae1365fa45ca7621bc050470e7e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67d090dbd4e672a6d3e777da244480eaf270518c68f8eaf2292ef7c366b8da4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-scwpp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:24Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.941579 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.941633 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.941649 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.941667 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:24 crc kubenswrapper[4873]: I1201 08:41:24.941683 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:24Z","lastTransitionTime":"2025-12-01T08:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.045993 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.046138 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.046160 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.046186 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.046202 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:25Z","lastTransitionTime":"2025-12-01T08:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.150435 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.150497 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.150512 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.150535 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.150551 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:25Z","lastTransitionTime":"2025-12-01T08:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.255252 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.255321 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.255342 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.255368 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.255387 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:25Z","lastTransitionTime":"2025-12-01T08:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.359069 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.359116 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.359128 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.359146 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.359158 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:25Z","lastTransitionTime":"2025-12-01T08:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.464333 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.464753 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.464770 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.464796 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.464816 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:25Z","lastTransitionTime":"2025-12-01T08:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.568574 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.568821 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.568838 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.568865 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.568883 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:25Z","lastTransitionTime":"2025-12-01T08:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.672207 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.672279 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.672298 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.672327 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.672346 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:25Z","lastTransitionTime":"2025-12-01T08:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.775379 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.775484 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.775505 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.775537 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.775555 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:25Z","lastTransitionTime":"2025-12-01T08:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.878475 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.879477 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.879684 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.879894 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.880225 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:25Z","lastTransitionTime":"2025-12-01T08:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.983347 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.983424 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.983443 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.983473 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:25 crc kubenswrapper[4873]: I1201 08:41:25.983493 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:25Z","lastTransitionTime":"2025-12-01T08:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.087372 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.087444 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.087463 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.087488 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.087509 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:26Z","lastTransitionTime":"2025-12-01T08:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.190649 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.190688 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.190699 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.190717 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.190727 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:26Z","lastTransitionTime":"2025-12-01T08:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.294598 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.294635 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.294645 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.294657 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.294666 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:26Z","lastTransitionTime":"2025-12-01T08:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.399988 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.400044 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.400053 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.400066 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.400075 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:26Z","lastTransitionTime":"2025-12-01T08:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.429089 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.429094 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.429243 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:26 crc kubenswrapper[4873]: E1201 08:41:26.429299 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:41:26 crc kubenswrapper[4873]: E1201 08:41:26.429456 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:26 crc kubenswrapper[4873]: E1201 08:41:26.429628 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.429094 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:26 crc kubenswrapper[4873]: E1201 08:41:26.430046 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.488786 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.489304 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.489442 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.489580 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.489704 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:26Z","lastTransitionTime":"2025-12-01T08:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:26 crc kubenswrapper[4873]: E1201 08:41:26.508256 4873 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963d6124-09f6-48b8-b38c-854877aa92e3\\\",\\\"systemUUID\\\":\\\"27c66b0b-2450-400c-b6ea-cdb9cbbf95ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:26Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.512769 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.512834 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.512851 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.512875 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.512894 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:26Z","lastTransitionTime":"2025-12-01T08:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:26 crc kubenswrapper[4873]: E1201 08:41:26.527497 4873 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963d6124-09f6-48b8-b38c-854877aa92e3\\\",\\\"systemUUID\\\":\\\"27c66b0b-2450-400c-b6ea-cdb9cbbf95ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:26Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.533314 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.533417 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.533433 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.533457 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.533475 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:26Z","lastTransitionTime":"2025-12-01T08:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:26 crc kubenswrapper[4873]: E1201 08:41:26.548917 4873 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963d6124-09f6-48b8-b38c-854877aa92e3\\\",\\\"systemUUID\\\":\\\"27c66b0b-2450-400c-b6ea-cdb9cbbf95ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:26Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.554167 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.554204 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.554219 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.554240 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.554254 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:26Z","lastTransitionTime":"2025-12-01T08:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:26 crc kubenswrapper[4873]: E1201 08:41:26.572309 4873 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963d6124-09f6-48b8-b38c-854877aa92e3\\\",\\\"systemUUID\\\":\\\"27c66b0b-2450-400c-b6ea-cdb9cbbf95ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:26Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.578149 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.578196 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.578206 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.578222 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.578235 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:26Z","lastTransitionTime":"2025-12-01T08:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:26 crc kubenswrapper[4873]: E1201 08:41:26.596866 4873 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963d6124-09f6-48b8-b38c-854877aa92e3\\\",\\\"systemUUID\\\":\\\"27c66b0b-2450-400c-b6ea-cdb9cbbf95ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:26Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:26 crc kubenswrapper[4873]: E1201 08:41:26.596971 4873 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.599575 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.599593 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.599602 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.599616 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.599627 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:26Z","lastTransitionTime":"2025-12-01T08:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.703140 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.703192 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.703201 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.703217 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.703272 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:26Z","lastTransitionTime":"2025-12-01T08:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.806779 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.806852 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.806872 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.806900 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.806919 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:26Z","lastTransitionTime":"2025-12-01T08:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.910053 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.910128 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.910169 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.910204 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:26 crc kubenswrapper[4873]: I1201 08:41:26.910229 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:26Z","lastTransitionTime":"2025-12-01T08:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.014367 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.014926 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.015091 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.015276 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.015406 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:27Z","lastTransitionTime":"2025-12-01T08:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.118291 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.118352 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.118367 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.118391 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.118405 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:27Z","lastTransitionTime":"2025-12-01T08:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.221546 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.221608 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.221626 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.221650 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.221668 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:27Z","lastTransitionTime":"2025-12-01T08:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.325239 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.325328 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.325356 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.325389 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.325413 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:27Z","lastTransitionTime":"2025-12-01T08:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.429243 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.429803 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.429940 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.430118 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.430253 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:27Z","lastTransitionTime":"2025-12-01T08:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.534686 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.535194 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.535422 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.535594 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.535779 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:27Z","lastTransitionTime":"2025-12-01T08:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.639509 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.639557 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.639565 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.639582 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.639593 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:27Z","lastTransitionTime":"2025-12-01T08:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.742763 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.742828 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.742839 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.742888 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.742899 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:27Z","lastTransitionTime":"2025-12-01T08:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.846213 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.846265 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.846279 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.846299 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.846312 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:27Z","lastTransitionTime":"2025-12-01T08:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.949137 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.949185 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.949197 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.949214 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:27 crc kubenswrapper[4873]: I1201 08:41:27.949226 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:27Z","lastTransitionTime":"2025-12-01T08:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.052990 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.053076 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.053096 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.053119 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.053136 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:28Z","lastTransitionTime":"2025-12-01T08:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.156111 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.156174 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.156192 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.156217 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.156236 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:28Z","lastTransitionTime":"2025-12-01T08:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.258826 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.258917 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.258942 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.258970 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.258989 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:28Z","lastTransitionTime":"2025-12-01T08:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.362367 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.362437 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.362455 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.362483 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.362500 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:28Z","lastTransitionTime":"2025-12-01T08:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.429466 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:28 crc kubenswrapper[4873]: E1201 08:41:28.429644 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.429892 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:28 crc kubenswrapper[4873]: E1201 08:41:28.429989 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.430297 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:28 crc kubenswrapper[4873]: E1201 08:41:28.430400 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.433180 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:41:28 crc kubenswrapper[4873]: E1201 08:41:28.433322 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.465291 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.465325 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.465336 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.465351 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.465360 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:28Z","lastTransitionTime":"2025-12-01T08:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.567522 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.567562 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.567582 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.567606 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.567623 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:28Z","lastTransitionTime":"2025-12-01T08:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.670757 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.670827 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.670845 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.670869 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.670886 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:28Z","lastTransitionTime":"2025-12-01T08:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.773347 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.773390 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.773400 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.773416 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.773427 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:28Z","lastTransitionTime":"2025-12-01T08:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.876372 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.876413 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.876424 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.876439 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.876451 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:28Z","lastTransitionTime":"2025-12-01T08:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.978363 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.978405 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.978419 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.978432 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:28 crc kubenswrapper[4873]: I1201 08:41:28.978447 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:28Z","lastTransitionTime":"2025-12-01T08:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:29 crc kubenswrapper[4873]: I1201 08:41:29.082491 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:29 crc kubenswrapper[4873]: I1201 08:41:29.082530 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:29 crc kubenswrapper[4873]: I1201 08:41:29.082541 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:29 crc kubenswrapper[4873]: I1201 08:41:29.082559 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:29 crc kubenswrapper[4873]: I1201 08:41:29.082571 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:29Z","lastTransitionTime":"2025-12-01T08:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:29 crc kubenswrapper[4873]: I1201 08:41:29.184974 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:29 crc kubenswrapper[4873]: I1201 08:41:29.185006 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:29 crc kubenswrapper[4873]: I1201 08:41:29.185035 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:29 crc kubenswrapper[4873]: I1201 08:41:29.185050 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:29 crc kubenswrapper[4873]: I1201 08:41:29.185063 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:29Z","lastTransitionTime":"2025-12-01T08:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:29 crc kubenswrapper[4873]: I1201 08:41:29.288120 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:29 crc kubenswrapper[4873]: I1201 08:41:29.288189 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:29 crc kubenswrapper[4873]: I1201 08:41:29.288210 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:29 crc kubenswrapper[4873]: I1201 08:41:29.288240 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:29 crc kubenswrapper[4873]: I1201 08:41:29.288263 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:29Z","lastTransitionTime":"2025-12-01T08:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:29 crc kubenswrapper[4873]: I1201 08:41:29.391580 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:29 crc kubenswrapper[4873]: I1201 08:41:29.391665 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:29 crc kubenswrapper[4873]: I1201 08:41:29.391691 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:29 crc kubenswrapper[4873]: I1201 08:41:29.391724 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:29 crc kubenswrapper[4873]: I1201 08:41:29.391754 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:29Z","lastTransitionTime":"2025-12-01T08:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:29 crc kubenswrapper[4873]: I1201 08:41:29.494521 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:29 crc kubenswrapper[4873]: I1201 08:41:29.494576 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:29 crc kubenswrapper[4873]: I1201 08:41:29.494587 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:29 crc kubenswrapper[4873]: I1201 08:41:29.494606 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:29 crc kubenswrapper[4873]: I1201 08:41:29.494618 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:29Z","lastTransitionTime":"2025-12-01T08:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:29 crc kubenswrapper[4873]: I1201 08:41:29.596981 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:29 crc kubenswrapper[4873]: I1201 08:41:29.597004 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:29 crc kubenswrapper[4873]: I1201 08:41:29.597036 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:29 crc kubenswrapper[4873]: I1201 08:41:29.597051 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:29 crc kubenswrapper[4873]: I1201 08:41:29.597061 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:29Z","lastTransitionTime":"2025-12-01T08:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:29 crc kubenswrapper[4873]: I1201 08:41:29.698968 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:29 crc kubenswrapper[4873]: I1201 08:41:29.699003 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:29 crc kubenswrapper[4873]: I1201 08:41:29.699027 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:29 crc kubenswrapper[4873]: I1201 08:41:29.699044 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:29 crc kubenswrapper[4873]: I1201 08:41:29.699062 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:29Z","lastTransitionTime":"2025-12-01T08:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:29 crc kubenswrapper[4873]: I1201 08:41:29.802171 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:29 crc kubenswrapper[4873]: I1201 08:41:29.802233 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:29 crc kubenswrapper[4873]: I1201 08:41:29.802244 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:29 crc kubenswrapper[4873]: I1201 08:41:29.802259 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:29 crc kubenswrapper[4873]: I1201 08:41:29.802269 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:29Z","lastTransitionTime":"2025-12-01T08:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:29 crc kubenswrapper[4873]: I1201 08:41:29.904573 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:29 crc kubenswrapper[4873]: I1201 08:41:29.904660 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:29 crc kubenswrapper[4873]: I1201 08:41:29.904707 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:29 crc kubenswrapper[4873]: I1201 08:41:29.904729 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:29 crc kubenswrapper[4873]: I1201 08:41:29.904745 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:29Z","lastTransitionTime":"2025-12-01T08:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.006792 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.006868 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.006902 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.006935 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.006954 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:30Z","lastTransitionTime":"2025-12-01T08:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.109780 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.109841 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.109854 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.109871 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.109882 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:30Z","lastTransitionTime":"2025-12-01T08:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.212205 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.212298 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.212316 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.212338 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.212358 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:30Z","lastTransitionTime":"2025-12-01T08:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.315866 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.315940 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.315963 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.315990 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.316010 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:30Z","lastTransitionTime":"2025-12-01T08:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.420008 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.420103 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.420114 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.420136 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.420149 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:30Z","lastTransitionTime":"2025-12-01T08:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.429750 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.429762 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.429749 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:30 crc kubenswrapper[4873]: E1201 08:41:30.429914 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.429749 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:41:30 crc kubenswrapper[4873]: E1201 08:41:30.431431 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:30 crc kubenswrapper[4873]: E1201 08:41:30.431606 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:41:30 crc kubenswrapper[4873]: E1201 08:41:30.431699 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.522707 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.523053 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.523138 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.523211 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.523272 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:30Z","lastTransitionTime":"2025-12-01T08:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.626081 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.626116 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.626127 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.626146 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.626157 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:30Z","lastTransitionTime":"2025-12-01T08:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.709076 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/08d02e64-a000-4f11-837e-82ad06d925f5-metrics-certs\") pod \"network-metrics-daemon-9vq2r\" (UID: \"08d02e64-a000-4f11-837e-82ad06d925f5\") " pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:41:30 crc kubenswrapper[4873]: E1201 08:41:30.709214 4873 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 08:41:30 crc kubenswrapper[4873]: E1201 08:41:30.709278 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/08d02e64-a000-4f11-837e-82ad06d925f5-metrics-certs podName:08d02e64-a000-4f11-837e-82ad06d925f5 nodeName:}" failed. No retries permitted until 2025-12-01 08:42:02.709257111 +0000 UTC m=+98.611365650 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/08d02e64-a000-4f11-837e-82ad06d925f5-metrics-certs") pod "network-metrics-daemon-9vq2r" (UID: "08d02e64-a000-4f11-837e-82ad06d925f5") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.729004 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.729060 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.729069 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.729085 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.729097 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:30Z","lastTransitionTime":"2025-12-01T08:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.831405 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.831747 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.831918 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.832088 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.832236 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:30Z","lastTransitionTime":"2025-12-01T08:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.935132 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.935180 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.935189 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.935206 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.935216 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:30Z","lastTransitionTime":"2025-12-01T08:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.945535 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-tpwwr_13a2246b-93bb-4586-98ee-53fc84aaae02/kube-multus/0.log" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.945696 4873 generic.go:334] "Generic (PLEG): container finished" podID="13a2246b-93bb-4586-98ee-53fc84aaae02" containerID="3dfcc763fe6ab69cab09f09c92ad9a85a4ded49c1443904134f78d1ca0092dae" exitCode=1 Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.945741 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-tpwwr" event={"ID":"13a2246b-93bb-4586-98ee-53fc84aaae02","Type":"ContainerDied","Data":"3dfcc763fe6ab69cab09f09c92ad9a85a4ded49c1443904134f78d1ca0092dae"} Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.946442 4873 scope.go:117] "RemoveContainer" containerID="3dfcc763fe6ab69cab09f09c92ad9a85a4ded49c1443904134f78d1ca0092dae" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.959733 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f7ac0b4-d048-46e4-bcee-4a271503e46d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8208cd5fda4df2e3b5a8bd90a18c8aca62b16cc2c60e06f52411f24f3d60dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba3cfd8a611c687414d7ecd6b02b423b351d7ac2a4d041085d4ca3bb2b0709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ae6257c73f3632c499b5f107178f64a9cac2f4325c57b41aa75d6d4ea14dd12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5088b5bf7bc847f59ca9efbf400855a70b99f1bbbf207eb93576c3b3eabae5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:30Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.979685 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:30Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:30 crc kubenswrapper[4873]: I1201 08:41:30.991664 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36539974b64191f4ca087bacbd02333f90c89fc8273ab7285e9a1d840155707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:30Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.004779 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef7b114-0e07-402d-a37b-315c36011f4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26e4a6e0a8877a213e2e974d4f7a3ad84938ae1365fa45ca7621bc050470e7e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67d090dbd4e672a6d3e777da244480eaf270518c68f8eaf2292ef7c366b8da4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-scwpp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:31Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.020021 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfqcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06182c66-4be6-4e00-9a17-4917badfc997\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa8ef5ef07129ec4a7b4cf4738a01cd6444e9bf548134a9da493c7149301387\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfqcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:31Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.033346 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9vq2r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08d02e64-a000-4f11-837e-82ad06d925f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzwvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzwvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9vq2r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:31Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.038603 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.038643 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.038655 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.038672 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.038683 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:31Z","lastTransitionTime":"2025-12-01T08:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.057970 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a2348f-009e-4cc1-b591-e1c14ff0faf2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff98ccdc9925e44e62f0d6e826ea7926dfa2a7d83d33e475a092f128b7fe920a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4676d506268107b87ebe2e21ba6e65d959a23a4d6f38bf0de588bd9c5d15e207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4751109230b67ad8948a2de4794571581d662bfd8abc6dc5c198dbc411172860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3acd19dc33ad40d8857d28bc0ccee4c99a6927e92e5029873234ecb5dc6ed18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7242d1034bd1938cb0bd750cab74db52d454ddd8af75f1e9e0a3c007dcdddd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:31Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.075182 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88068a1cf89e2c507b9d93d305db6af5af196bd9eec805f42d91b6fe090aba35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:31Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.092261 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:31Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.107311 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:31Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.125914 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a308045b-f54d-42ec-8fdf-5bc3e54ef363\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d41610d208e2751d8f43b01b041454558f7764717c0855470dd5680a61a34e6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d41610d208e2751d8f43b01b041454558f7764717c0855470dd5680a61a34e6f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"message\\\":\\\" lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-image-registry/image-registry_TCP_cluster\\\\\\\", UUID:\\\\\\\"83c1e277-3d22-42ae-a355-f7a0ff0bd171\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-image-registry/image-registry\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-image-registry/image-registry_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-image-registry/image-registry\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.93\\\\\\\", Port:5000, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1201 08:41:10.392448 6540 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:41:09Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dxlvk_openshift-ovn-kubernetes(a308045b-f54d-42ec-8fdf-5bc3e54ef363)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dxlvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:31Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.141858 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.141886 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.141894 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.141911 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.141924 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:31Z","lastTransitionTime":"2025-12-01T08:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.144731 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5smrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdde424-6124-432b-9902-cedb9dca26c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6c098fa4f45096e671331702f2995119b74cc8c87c5fa1b175099070bc9099\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lvr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cfdfa53bea0d7e308f7673baf05e104ad7967e0473e36a74a17825433a17c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lvr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5smrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:31Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.156984 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6938caf5-0969-4f3a-ba82-ae50109d7be0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25326d274f24d0da05778ed8c5a80d465065df87cf625be2c35d9c95763d4aab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9ffbc1a66dd17bac0e36d480467c72521d7a97c5376dafe57fa15efe571c78b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ce4df67c597a961222280dff543040c444f4d7eace37c5cfe07c7f9d89ea70a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aae94f574fbe3c1c7fcd3dbe889f7aa484e1482cee011ebfb2044151d4ca8eac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aae94f574fbe3c1c7fcd3dbe889f7aa484e1482cee011ebfb2044151d4ca8eac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:31Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.169477 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1c8f149-f188-4759-b2dc-9dc15be1bb13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:40:37.024907 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:40:37.026665 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1609525021/tls.crt::/tmp/serving-cert-1609525021/tls.key\\\\\\\"\\\\nI1201 08:40:42.169070 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:40:42.177896 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:40:42.177940 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:40:42.177968 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:40:42.177975 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:40:42.184767 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1201 08:40:42.184781 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1201 08:40:42.184802 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184808 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184814 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:40:42.184819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:40:42.184822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:40:42.184826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1201 08:40:42.187129 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:31Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.187406 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tpwwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13a2246b-93bb-4586-98ee-53fc84aaae02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dfcc763fe6ab69cab09f09c92ad9a85a4ded49c1443904134f78d1ca0092dae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dfcc763fe6ab69cab09f09c92ad9a85a4ded49c1443904134f78d1ca0092dae\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:41:30Z\\\",\\\"message\\\":\\\"2025-12-01T08:40:45+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_81306c4a-540c-48bc-b1a5-86f626f67123\\\\n2025-12-01T08:40:45+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_81306c4a-540c-48bc-b1a5-86f626f67123 to /host/opt/cni/bin/\\\\n2025-12-01T08:40:45Z [verbose] multus-daemon started\\\\n2025-12-01T08:40:45Z [verbose] Readiness Indicator file check\\\\n2025-12-01T08:41:30Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zfr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tpwwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:31Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.201216 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90f08b6fa658ce25fbc25d8d44d956cdb6424a0b047205f19d1a6fd6870b1e90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8d27e3678f1f63f55bd1b76668e328f49368d67d4d7338a6326cb73d13e08f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:31Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.211732 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qpr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c42ab88e96b59d9430a69343e06a329447ae0163cb0f5b9ef2bcafa1bf356583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qpr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:31Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.229344 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8570b876-a13b-4b8f-99c4-4900cfe19a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0afb7d55f395ee39aa245516631b88aab92399973a628f91172c9e410dcc13dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shkn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:31Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.245660 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.245704 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.245713 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.245732 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.245743 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:31Z","lastTransitionTime":"2025-12-01T08:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.347936 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.347991 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.348004 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.348039 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.348049 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:31Z","lastTransitionTime":"2025-12-01T08:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.450534 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.450570 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.450581 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.450596 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.450609 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:31Z","lastTransitionTime":"2025-12-01T08:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.552919 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.552996 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.553061 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.553099 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.553130 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:31Z","lastTransitionTime":"2025-12-01T08:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.656916 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.657559 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.657577 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.657602 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.657618 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:31Z","lastTransitionTime":"2025-12-01T08:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.761537 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.761629 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.761653 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.761685 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.761708 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:31Z","lastTransitionTime":"2025-12-01T08:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.865390 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.865478 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.865503 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.865533 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.865556 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:31Z","lastTransitionTime":"2025-12-01T08:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.952307 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-tpwwr_13a2246b-93bb-4586-98ee-53fc84aaae02/kube-multus/0.log" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.952371 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-tpwwr" event={"ID":"13a2246b-93bb-4586-98ee-53fc84aaae02","Type":"ContainerStarted","Data":"a9f46a2651e067cd11996fb1c43221a21ab71fdc3ca78a61d9245516f871a5f3"} Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.966736 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90f08b6fa658ce25fbc25d8d44d956cdb6424a0b047205f19d1a6fd6870b1e90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8d27e3678f1f63f55bd1b76668e328f49368d67d4d7338a6326cb73d13e08f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:31Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.968058 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.968272 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.968806 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.968986 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.969305 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:31Z","lastTransitionTime":"2025-12-01T08:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.979190 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qpr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c42ab88e96b59d9430a69343e06a329447ae0163cb0f5b9ef2bcafa1bf356583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qpr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:31Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:31 crc kubenswrapper[4873]: I1201 08:41:31.995700 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8570b876-a13b-4b8f-99c4-4900cfe19a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0afb7d55f395ee39aa245516631b88aab92399973a628f91172c9e410dcc13dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shkn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:31Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.007935 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f7ac0b4-d048-46e4-bcee-4a271503e46d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8208cd5fda4df2e3b5a8bd90a18c8aca62b16cc2c60e06f52411f24f3d60dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba3cfd8a611c687414d7ecd6b02b423b351d7ac2a4d041085d4ca3bb2b0709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ae6257c73f3632c499b5f107178f64a9cac2f4325c57b41aa75d6d4ea14dd12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5088b5bf7bc847f59ca9efbf400855a70b99f1bbbf207eb93576c3b3eabae5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:32Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.023089 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:32Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.035770 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36539974b64191f4ca087bacbd02333f90c89fc8273ab7285e9a1d840155707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:32Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.045892 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef7b114-0e07-402d-a37b-315c36011f4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26e4a6e0a8877a213e2e974d4f7a3ad84938ae1365fa45ca7621bc050470e7e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67d090dbd4e672a6d3e777da244480eaf270518c68f8eaf2292ef7c366b8da4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-scwpp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:32Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.057670 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfqcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06182c66-4be6-4e00-9a17-4917badfc997\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa8ef5ef07129ec4a7b4cf4738a01cd6444e9bf548134a9da493c7149301387\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfqcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:32Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.067060 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9vq2r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08d02e64-a000-4f11-837e-82ad06d925f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzwvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzwvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9vq2r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:32Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.072025 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.072086 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.072099 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.072121 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.072132 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:32Z","lastTransitionTime":"2025-12-01T08:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.084585 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a2348f-009e-4cc1-b591-e1c14ff0faf2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff98ccdc9925e44e62f0d6e826ea7926dfa2a7d83d33e475a092f128b7fe920a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4676d506268107b87ebe2e21ba6e65d959a23a4d6f38bf0de588bd9c5d15e207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4751109230b67ad8948a2de4794571581d662bfd8abc6dc5c198dbc411172860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3acd19dc33ad40d8857d28bc0ccee4c99a6927e92e5029873234ecb5dc6ed18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7242d1034bd1938cb0bd750cab74db52d454ddd8af75f1e9e0a3c007dcdddd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:32Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.098630 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88068a1cf89e2c507b9d93d305db6af5af196bd9eec805f42d91b6fe090aba35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:32Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.110114 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:32Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.126042 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:32Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.151774 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a308045b-f54d-42ec-8fdf-5bc3e54ef363\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d41610d208e2751d8f43b01b041454558f7764717c0855470dd5680a61a34e6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d41610d208e2751d8f43b01b041454558f7764717c0855470dd5680a61a34e6f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"message\\\":\\\" lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-image-registry/image-registry_TCP_cluster\\\\\\\", UUID:\\\\\\\"83c1e277-3d22-42ae-a355-f7a0ff0bd171\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-image-registry/image-registry\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-image-registry/image-registry_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-image-registry/image-registry\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.93\\\\\\\", Port:5000, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1201 08:41:10.392448 6540 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:41:09Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dxlvk_openshift-ovn-kubernetes(a308045b-f54d-42ec-8fdf-5bc3e54ef363)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dxlvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:32Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.166811 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5smrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdde424-6124-432b-9902-cedb9dca26c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6c098fa4f45096e671331702f2995119b74cc8c87c5fa1b175099070bc9099\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lvr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cfdfa53bea0d7e308f7673baf05e104ad7967e0473e36a74a17825433a17c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lvr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5smrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:32Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.177913 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.177957 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.177968 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.177989 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.178004 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:32Z","lastTransitionTime":"2025-12-01T08:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.180831 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6938caf5-0969-4f3a-ba82-ae50109d7be0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25326d274f24d0da05778ed8c5a80d465065df87cf625be2c35d9c95763d4aab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9ffbc1a66dd17bac0e36d480467c72521d7a97c5376dafe57fa15efe571c78b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ce4df67c597a961222280dff543040c444f4d7eace37c5cfe07c7f9d89ea70a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aae94f574fbe3c1c7fcd3dbe889f7aa484e1482cee011ebfb2044151d4ca8eac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aae94f574fbe3c1c7fcd3dbe889f7aa484e1482cee011ebfb2044151d4ca8eac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:32Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.200434 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1c8f149-f188-4759-b2dc-9dc15be1bb13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:40:37.024907 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:40:37.026665 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1609525021/tls.crt::/tmp/serving-cert-1609525021/tls.key\\\\\\\"\\\\nI1201 08:40:42.169070 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:40:42.177896 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:40:42.177940 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:40:42.177968 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:40:42.177975 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:40:42.184767 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1201 08:40:42.184781 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1201 08:40:42.184802 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184808 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184814 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:40:42.184819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:40:42.184822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:40:42.184826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1201 08:40:42.187129 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:32Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.215252 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tpwwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13a2246b-93bb-4586-98ee-53fc84aaae02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9f46a2651e067cd11996fb1c43221a21ab71fdc3ca78a61d9245516f871a5f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dfcc763fe6ab69cab09f09c92ad9a85a4ded49c1443904134f78d1ca0092dae\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:41:30Z\\\",\\\"message\\\":\\\"2025-12-01T08:40:45+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_81306c4a-540c-48bc-b1a5-86f626f67123\\\\n2025-12-01T08:40:45+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_81306c4a-540c-48bc-b1a5-86f626f67123 to /host/opt/cni/bin/\\\\n2025-12-01T08:40:45Z [verbose] multus-daemon started\\\\n2025-12-01T08:40:45Z [verbose] Readiness Indicator file check\\\\n2025-12-01T08:41:30Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zfr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tpwwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:32Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.280668 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.281117 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.281225 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.281309 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.281376 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:32Z","lastTransitionTime":"2025-12-01T08:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.384836 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.384899 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.384911 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.384938 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.384949 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:32Z","lastTransitionTime":"2025-12-01T08:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.429930 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.430033 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.429927 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:32 crc kubenswrapper[4873]: E1201 08:41:32.430138 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.430227 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:41:32 crc kubenswrapper[4873]: E1201 08:41:32.430395 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:32 crc kubenswrapper[4873]: E1201 08:41:32.430557 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:32 crc kubenswrapper[4873]: E1201 08:41:32.430744 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.487548 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.487662 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.487689 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.487717 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.487737 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:32Z","lastTransitionTime":"2025-12-01T08:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.590780 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.590854 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.590880 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.590913 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.591060 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:32Z","lastTransitionTime":"2025-12-01T08:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.694835 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.695210 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.695371 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.695519 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.695654 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:32Z","lastTransitionTime":"2025-12-01T08:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.798711 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.798775 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.798794 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.798825 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.798843 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:32Z","lastTransitionTime":"2025-12-01T08:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.901306 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.901713 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.901881 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.902113 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:32 crc kubenswrapper[4873]: I1201 08:41:32.902383 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:32Z","lastTransitionTime":"2025-12-01T08:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.004480 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.004974 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.005162 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.005311 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.005461 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:33Z","lastTransitionTime":"2025-12-01T08:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.109811 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.109914 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.109933 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.109961 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.109983 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:33Z","lastTransitionTime":"2025-12-01T08:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.213083 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.213126 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.213137 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.213154 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.213165 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:33Z","lastTransitionTime":"2025-12-01T08:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.315630 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.315676 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.315688 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.315707 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.315718 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:33Z","lastTransitionTime":"2025-12-01T08:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.418723 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.418779 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.418792 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.418810 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.418825 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:33Z","lastTransitionTime":"2025-12-01T08:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.429550 4873 scope.go:117] "RemoveContainer" containerID="d41610d208e2751d8f43b01b041454558f7764717c0855470dd5680a61a34e6f" Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.524585 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.524629 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.524640 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.524658 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.524670 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:33Z","lastTransitionTime":"2025-12-01T08:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.627469 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.627501 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.627511 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.627529 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.627540 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:33Z","lastTransitionTime":"2025-12-01T08:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.730137 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.730176 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.730186 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.730201 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.730212 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:33Z","lastTransitionTime":"2025-12-01T08:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.833094 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.833128 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.833138 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.833155 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.833167 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:33Z","lastTransitionTime":"2025-12-01T08:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.934871 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.934942 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.934955 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.934974 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.934984 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:33Z","lastTransitionTime":"2025-12-01T08:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.958907 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dxlvk_a308045b-f54d-42ec-8fdf-5bc3e54ef363/ovnkube-controller/2.log" Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.960966 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" event={"ID":"a308045b-f54d-42ec-8fdf-5bc3e54ef363","Type":"ContainerStarted","Data":"c55ab10131e65cc646cead8704fd95f3534c6726b01283d2e3fdf4b8da4e03ff"} Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.961414 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.976503 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5smrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdde424-6124-432b-9902-cedb9dca26c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6c098fa4f45096e671331702f2995119b74cc8c87c5fa1b175099070bc9099\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lvr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cfdfa53bea0d7e308f7673baf05e104ad7967e0473e36a74a17825433a17c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lvr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5smrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:33Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:33 crc kubenswrapper[4873]: I1201 08:41:33.988957 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9vq2r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08d02e64-a000-4f11-837e-82ad06d925f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzwvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzwvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9vq2r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:33Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.009797 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a2348f-009e-4cc1-b591-e1c14ff0faf2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff98ccdc9925e44e62f0d6e826ea7926dfa2a7d83d33e475a092f128b7fe920a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4676d506268107b87ebe2e21ba6e65d959a23a4d6f38bf0de588bd9c5d15e207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4751109230b67ad8948a2de4794571581d662bfd8abc6dc5c198dbc411172860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3acd19dc33ad40d8857d28bc0ccee4c99a6927e92e5029873234ecb5dc6ed18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7242d1034bd1938cb0bd750cab74db52d454ddd8af75f1e9e0a3c007dcdddd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:34Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.037041 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.037098 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.037110 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.037126 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.037138 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:34Z","lastTransitionTime":"2025-12-01T08:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.049962 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88068a1cf89e2c507b9d93d305db6af5af196bd9eec805f42d91b6fe090aba35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:34Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.078315 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:34Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.098606 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:34Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.121387 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a308045b-f54d-42ec-8fdf-5bc3e54ef363\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55ab10131e65cc646cead8704fd95f3534c6726b01283d2e3fdf4b8da4e03ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d41610d208e2751d8f43b01b041454558f7764717c0855470dd5680a61a34e6f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"message\\\":\\\" lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-image-registry/image-registry_TCP_cluster\\\\\\\", UUID:\\\\\\\"83c1e277-3d22-42ae-a355-f7a0ff0bd171\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-image-registry/image-registry\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-image-registry/image-registry_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-image-registry/image-registry\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.93\\\\\\\", Port:5000, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1201 08:41:10.392448 6540 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:41:09Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:41:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dxlvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:34Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.135189 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6938caf5-0969-4f3a-ba82-ae50109d7be0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25326d274f24d0da05778ed8c5a80d465065df87cf625be2c35d9c95763d4aab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9ffbc1a66dd17bac0e36d480467c72521d7a97c5376dafe57fa15efe571c78b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ce4df67c597a961222280dff543040c444f4d7eace37c5cfe07c7f9d89ea70a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aae94f574fbe3c1c7fcd3dbe889f7aa484e1482cee011ebfb2044151d4ca8eac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aae94f574fbe3c1c7fcd3dbe889f7aa484e1482cee011ebfb2044151d4ca8eac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:34Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.139119 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.139163 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.139172 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.139191 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.139204 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:34Z","lastTransitionTime":"2025-12-01T08:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.152379 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1c8f149-f188-4759-b2dc-9dc15be1bb13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:40:37.024907 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:40:37.026665 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1609525021/tls.crt::/tmp/serving-cert-1609525021/tls.key\\\\\\\"\\\\nI1201 08:40:42.169070 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:40:42.177896 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:40:42.177940 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:40:42.177968 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:40:42.177975 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:40:42.184767 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1201 08:40:42.184781 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1201 08:40:42.184802 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184808 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184814 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:40:42.184819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:40:42.184822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:40:42.184826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1201 08:40:42.187129 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:34Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.171103 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tpwwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13a2246b-93bb-4586-98ee-53fc84aaae02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9f46a2651e067cd11996fb1c43221a21ab71fdc3ca78a61d9245516f871a5f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dfcc763fe6ab69cab09f09c92ad9a85a4ded49c1443904134f78d1ca0092dae\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:41:30Z\\\",\\\"message\\\":\\\"2025-12-01T08:40:45+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_81306c4a-540c-48bc-b1a5-86f626f67123\\\\n2025-12-01T08:40:45+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_81306c4a-540c-48bc-b1a5-86f626f67123 to /host/opt/cni/bin/\\\\n2025-12-01T08:40:45Z [verbose] multus-daemon started\\\\n2025-12-01T08:40:45Z [verbose] Readiness Indicator file check\\\\n2025-12-01T08:41:30Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zfr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tpwwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:34Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.185832 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90f08b6fa658ce25fbc25d8d44d956cdb6424a0b047205f19d1a6fd6870b1e90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8d27e3678f1f63f55bd1b76668e328f49368d67d4d7338a6326cb73d13e08f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:34Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.196848 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qpr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c42ab88e96b59d9430a69343e06a329447ae0163cb0f5b9ef2bcafa1bf356583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qpr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:34Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.214878 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8570b876-a13b-4b8f-99c4-4900cfe19a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0afb7d55f395ee39aa245516631b88aab92399973a628f91172c9e410dcc13dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shkn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:34Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.228998 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f7ac0b4-d048-46e4-bcee-4a271503e46d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8208cd5fda4df2e3b5a8bd90a18c8aca62b16cc2c60e06f52411f24f3d60dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba3cfd8a611c687414d7ecd6b02b423b351d7ac2a4d041085d4ca3bb2b0709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ae6257c73f3632c499b5f107178f64a9cac2f4325c57b41aa75d6d4ea14dd12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5088b5bf7bc847f59ca9efbf400855a70b99f1bbbf207eb93576c3b3eabae5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:34Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.241468 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.241503 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.241513 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.241527 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.241542 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:34Z","lastTransitionTime":"2025-12-01T08:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.250385 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:34Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.266315 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36539974b64191f4ca087bacbd02333f90c89fc8273ab7285e9a1d840155707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:34Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.281455 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef7b114-0e07-402d-a37b-315c36011f4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26e4a6e0a8877a213e2e974d4f7a3ad84938ae1365fa45ca7621bc050470e7e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67d090dbd4e672a6d3e777da244480eaf270518c68f8eaf2292ef7c366b8da4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-scwpp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:34Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.292557 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfqcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06182c66-4be6-4e00-9a17-4917badfc997\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa8ef5ef07129ec4a7b4cf4738a01cd6444e9bf548134a9da493c7149301387\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfqcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:34Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.343652 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.343688 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.343709 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.343726 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.343741 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:34Z","lastTransitionTime":"2025-12-01T08:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.429874 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.429923 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.429970 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.430033 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:41:34 crc kubenswrapper[4873]: E1201 08:41:34.430155 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:34 crc kubenswrapper[4873]: E1201 08:41:34.430324 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:41:34 crc kubenswrapper[4873]: E1201 08:41:34.430514 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:34 crc kubenswrapper[4873]: E1201 08:41:34.430627 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.442723 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88068a1cf89e2c507b9d93d305db6af5af196bd9eec805f42d91b6fe090aba35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:34Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.445843 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.445893 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.445906 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.445924 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.445937 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:34Z","lastTransitionTime":"2025-12-01T08:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.454654 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:34Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.472801 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:34Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.491629 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a308045b-f54d-42ec-8fdf-5bc3e54ef363\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55ab10131e65cc646cead8704fd95f3534c6726b01283d2e3fdf4b8da4e03ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d41610d208e2751d8f43b01b041454558f7764717c0855470dd5680a61a34e6f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"message\\\":\\\" lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-image-registry/image-registry_TCP_cluster\\\\\\\", UUID:\\\\\\\"83c1e277-3d22-42ae-a355-f7a0ff0bd171\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-image-registry/image-registry\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-image-registry/image-registry_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-image-registry/image-registry\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.93\\\\\\\", Port:5000, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1201 08:41:10.392448 6540 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:41:09Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:41:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dxlvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:34Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.501987 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5smrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdde424-6124-432b-9902-cedb9dca26c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6c098fa4f45096e671331702f2995119b74cc8c87c5fa1b175099070bc9099\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lvr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cfdfa53bea0d7e308f7673baf05e104ad7967e0473e36a74a17825433a17c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lvr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5smrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:34Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.515212 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9vq2r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08d02e64-a000-4f11-837e-82ad06d925f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzwvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzwvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9vq2r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:34Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.544352 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a2348f-009e-4cc1-b591-e1c14ff0faf2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff98ccdc9925e44e62f0d6e826ea7926dfa2a7d83d33e475a092f128b7fe920a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4676d506268107b87ebe2e21ba6e65d959a23a4d6f38bf0de588bd9c5d15e207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4751109230b67ad8948a2de4794571581d662bfd8abc6dc5c198dbc411172860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3acd19dc33ad40d8857d28bc0ccee4c99a6927e92e5029873234ecb5dc6ed18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7242d1034bd1938cb0bd750cab74db52d454ddd8af75f1e9e0a3c007dcdddd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:34Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.548438 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.548487 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.548499 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.548520 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.548534 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:34Z","lastTransitionTime":"2025-12-01T08:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.562145 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1c8f149-f188-4759-b2dc-9dc15be1bb13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:40:37.024907 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:40:37.026665 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1609525021/tls.crt::/tmp/serving-cert-1609525021/tls.key\\\\\\\"\\\\nI1201 08:40:42.169070 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:40:42.177896 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:40:42.177940 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:40:42.177968 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:40:42.177975 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:40:42.184767 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1201 08:40:42.184781 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1201 08:40:42.184802 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184808 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184814 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:40:42.184819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:40:42.184822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:40:42.184826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1201 08:40:42.187129 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:34Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.578836 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tpwwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13a2246b-93bb-4586-98ee-53fc84aaae02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9f46a2651e067cd11996fb1c43221a21ab71fdc3ca78a61d9245516f871a5f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dfcc763fe6ab69cab09f09c92ad9a85a4ded49c1443904134f78d1ca0092dae\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:41:30Z\\\",\\\"message\\\":\\\"2025-12-01T08:40:45+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_81306c4a-540c-48bc-b1a5-86f626f67123\\\\n2025-12-01T08:40:45+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_81306c4a-540c-48bc-b1a5-86f626f67123 to /host/opt/cni/bin/\\\\n2025-12-01T08:40:45Z [verbose] multus-daemon started\\\\n2025-12-01T08:40:45Z [verbose] Readiness Indicator file check\\\\n2025-12-01T08:41:30Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zfr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tpwwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:34Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.596011 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6938caf5-0969-4f3a-ba82-ae50109d7be0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25326d274f24d0da05778ed8c5a80d465065df87cf625be2c35d9c95763d4aab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9ffbc1a66dd17bac0e36d480467c72521d7a97c5376dafe57fa15efe571c78b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ce4df67c597a961222280dff543040c444f4d7eace37c5cfe07c7f9d89ea70a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aae94f574fbe3c1c7fcd3dbe889f7aa484e1482cee011ebfb2044151d4ca8eac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aae94f574fbe3c1c7fcd3dbe889f7aa484e1482cee011ebfb2044151d4ca8eac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:34Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.606755 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qpr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c42ab88e96b59d9430a69343e06a329447ae0163cb0f5b9ef2bcafa1bf356583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qpr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:34Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.622327 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8570b876-a13b-4b8f-99c4-4900cfe19a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0afb7d55f395ee39aa245516631b88aab92399973a628f91172c9e410dcc13dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shkn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:34Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.636183 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90f08b6fa658ce25fbc25d8d44d956cdb6424a0b047205f19d1a6fd6870b1e90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8d27e3678f1f63f55bd1b76668e328f49368d67d4d7338a6326cb73d13e08f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:34Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.650951 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:34Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.651685 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.651735 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.651747 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.651769 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.651787 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:34Z","lastTransitionTime":"2025-12-01T08:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.664197 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36539974b64191f4ca087bacbd02333f90c89fc8273ab7285e9a1d840155707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:34Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.677460 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef7b114-0e07-402d-a37b-315c36011f4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26e4a6e0a8877a213e2e974d4f7a3ad84938ae1365fa45ca7621bc050470e7e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67d090dbd4e672a6d3e777da244480eaf270518c68f8eaf2292ef7c366b8da4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-scwpp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:34Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.689510 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfqcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06182c66-4be6-4e00-9a17-4917badfc997\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa8ef5ef07129ec4a7b4cf4738a01cd6444e9bf548134a9da493c7149301387\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfqcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:34Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.703778 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f7ac0b4-d048-46e4-bcee-4a271503e46d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8208cd5fda4df2e3b5a8bd90a18c8aca62b16cc2c60e06f52411f24f3d60dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba3cfd8a611c687414d7ecd6b02b423b351d7ac2a4d041085d4ca3bb2b0709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ae6257c73f3632c499b5f107178f64a9cac2f4325c57b41aa75d6d4ea14dd12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5088b5bf7bc847f59ca9efbf400855a70b99f1bbbf207eb93576c3b3eabae5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:34Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.755206 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.755285 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.755305 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.755332 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.755352 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:34Z","lastTransitionTime":"2025-12-01T08:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.859191 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.859243 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.859256 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.859273 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.859283 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:34Z","lastTransitionTime":"2025-12-01T08:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.962548 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.962597 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.962610 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.962626 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.962639 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:34Z","lastTransitionTime":"2025-12-01T08:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.965754 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dxlvk_a308045b-f54d-42ec-8fdf-5bc3e54ef363/ovnkube-controller/3.log" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.966504 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dxlvk_a308045b-f54d-42ec-8fdf-5bc3e54ef363/ovnkube-controller/2.log" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.969637 4873 generic.go:334] "Generic (PLEG): container finished" podID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerID="c55ab10131e65cc646cead8704fd95f3534c6726b01283d2e3fdf4b8da4e03ff" exitCode=1 Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.969698 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" event={"ID":"a308045b-f54d-42ec-8fdf-5bc3e54ef363","Type":"ContainerDied","Data":"c55ab10131e65cc646cead8704fd95f3534c6726b01283d2e3fdf4b8da4e03ff"} Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.969749 4873 scope.go:117] "RemoveContainer" containerID="d41610d208e2751d8f43b01b041454558f7764717c0855470dd5680a61a34e6f" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.970664 4873 scope.go:117] "RemoveContainer" containerID="c55ab10131e65cc646cead8704fd95f3534c6726b01283d2e3fdf4b8da4e03ff" Dec 01 08:41:34 crc kubenswrapper[4873]: E1201 08:41:34.970964 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dxlvk_openshift-ovn-kubernetes(a308045b-f54d-42ec-8fdf-5bc3e54ef363)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" Dec 01 08:41:34 crc kubenswrapper[4873]: I1201 08:41:34.988971 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6938caf5-0969-4f3a-ba82-ae50109d7be0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25326d274f24d0da05778ed8c5a80d465065df87cf625be2c35d9c95763d4aab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9ffbc1a66dd17bac0e36d480467c72521d7a97c5376dafe57fa15efe571c78b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ce4df67c597a961222280dff543040c444f4d7eace37c5cfe07c7f9d89ea70a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aae94f574fbe3c1c7fcd3dbe889f7aa484e1482cee011ebfb2044151d4ca8eac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aae94f574fbe3c1c7fcd3dbe889f7aa484e1482cee011ebfb2044151d4ca8eac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:34Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.010264 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1c8f149-f188-4759-b2dc-9dc15be1bb13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:40:37.024907 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:40:37.026665 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1609525021/tls.crt::/tmp/serving-cert-1609525021/tls.key\\\\\\\"\\\\nI1201 08:40:42.169070 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:40:42.177896 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:40:42.177940 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:40:42.177968 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:40:42.177975 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:40:42.184767 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1201 08:40:42.184781 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1201 08:40:42.184802 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184808 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184814 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:40:42.184819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:40:42.184822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:40:42.184826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1201 08:40:42.187129 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:35Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.029152 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tpwwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13a2246b-93bb-4586-98ee-53fc84aaae02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9f46a2651e067cd11996fb1c43221a21ab71fdc3ca78a61d9245516f871a5f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dfcc763fe6ab69cab09f09c92ad9a85a4ded49c1443904134f78d1ca0092dae\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:41:30Z\\\",\\\"message\\\":\\\"2025-12-01T08:40:45+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_81306c4a-540c-48bc-b1a5-86f626f67123\\\\n2025-12-01T08:40:45+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_81306c4a-540c-48bc-b1a5-86f626f67123 to /host/opt/cni/bin/\\\\n2025-12-01T08:40:45Z [verbose] multus-daemon started\\\\n2025-12-01T08:40:45Z [verbose] Readiness Indicator file check\\\\n2025-12-01T08:41:30Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zfr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tpwwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:35Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.046433 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90f08b6fa658ce25fbc25d8d44d956cdb6424a0b047205f19d1a6fd6870b1e90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8d27e3678f1f63f55bd1b76668e328f49368d67d4d7338a6326cb73d13e08f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:35Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.059362 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qpr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c42ab88e96b59d9430a69343e06a329447ae0163cb0f5b9ef2bcafa1bf356583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qpr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:35Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.064692 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.064753 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.064764 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.064800 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.064812 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:35Z","lastTransitionTime":"2025-12-01T08:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.077416 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8570b876-a13b-4b8f-99c4-4900cfe19a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0afb7d55f395ee39aa245516631b88aab92399973a628f91172c9e410dcc13dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shkn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:35Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.096807 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f7ac0b4-d048-46e4-bcee-4a271503e46d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8208cd5fda4df2e3b5a8bd90a18c8aca62b16cc2c60e06f52411f24f3d60dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba3cfd8a611c687414d7ecd6b02b423b351d7ac2a4d041085d4ca3bb2b0709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ae6257c73f3632c499b5f107178f64a9cac2f4325c57b41aa75d6d4ea14dd12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5088b5bf7bc847f59ca9efbf400855a70b99f1bbbf207eb93576c3b3eabae5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:35Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.117257 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:35Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.133180 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36539974b64191f4ca087bacbd02333f90c89fc8273ab7285e9a1d840155707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:35Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.149266 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef7b114-0e07-402d-a37b-315c36011f4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26e4a6e0a8877a213e2e974d4f7a3ad84938ae1365fa45ca7621bc050470e7e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67d090dbd4e672a6d3e777da244480eaf270518c68f8eaf2292ef7c366b8da4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-scwpp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:35Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.164908 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfqcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06182c66-4be6-4e00-9a17-4917badfc997\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa8ef5ef07129ec4a7b4cf4738a01cd6444e9bf548134a9da493c7149301387\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfqcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:35Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.167642 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.167690 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.167699 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.167713 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.167725 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:35Z","lastTransitionTime":"2025-12-01T08:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.184855 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5smrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdde424-6124-432b-9902-cedb9dca26c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6c098fa4f45096e671331702f2995119b74cc8c87c5fa1b175099070bc9099\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lvr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cfdfa53bea0d7e308f7673baf05e104ad7967e0473e36a74a17825433a17c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lvr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5smrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:35Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.200527 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9vq2r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08d02e64-a000-4f11-837e-82ad06d925f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzwvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzwvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9vq2r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:35Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.236839 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a2348f-009e-4cc1-b591-e1c14ff0faf2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff98ccdc9925e44e62f0d6e826ea7926dfa2a7d83d33e475a092f128b7fe920a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4676d506268107b87ebe2e21ba6e65d959a23a4d6f38bf0de588bd9c5d15e207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4751109230b67ad8948a2de4794571581d662bfd8abc6dc5c198dbc411172860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3acd19dc33ad40d8857d28bc0ccee4c99a6927e92e5029873234ecb5dc6ed18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7242d1034bd1938cb0bd750cab74db52d454ddd8af75f1e9e0a3c007dcdddd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:35Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.253123 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88068a1cf89e2c507b9d93d305db6af5af196bd9eec805f42d91b6fe090aba35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:35Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.267079 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:35Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.272246 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.272324 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.272337 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.272358 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.272393 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:35Z","lastTransitionTime":"2025-12-01T08:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.288669 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:35Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.309710 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a308045b-f54d-42ec-8fdf-5bc3e54ef363\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55ab10131e65cc646cead8704fd95f3534c6726b01283d2e3fdf4b8da4e03ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d41610d208e2751d8f43b01b041454558f7764717c0855470dd5680a61a34e6f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"message\\\":\\\" lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-image-registry/image-registry_TCP_cluster\\\\\\\", UUID:\\\\\\\"83c1e277-3d22-42ae-a355-f7a0ff0bd171\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-image-registry/image-registry\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-image-registry/image-registry_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-image-registry/image-registry\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.93\\\\\\\", Port:5000, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1201 08:41:10.392448 6540 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:41:09Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c55ab10131e65cc646cead8704fd95f3534c6726b01283d2e3fdf4b8da4e03ff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:41:34Z\\\",\\\"message\\\":\\\"Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5smrf after 0 failed attempt(s)\\\\nI1201 08:41:34.466741 6856 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5smrf\\\\nI1201 08:41:34.466681 6856 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-shkn9\\\\nI1201 08:41:34.466583 6856 default_network_controller.go:776] Recording success event on pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1201 08:41:34.466600 6856 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nI1201 08:41:34.466782 6856 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-qpr6r in node crc\\\\nI1201 08:41:34.466816 6856 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI1201 08:41:34.466591 6856 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nF1201 08:41:34.466439 6856 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:41:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dxlvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:35Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.375090 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.375138 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.375148 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.375166 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.375175 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:35Z","lastTransitionTime":"2025-12-01T08:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.477074 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.477123 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.477135 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.477154 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.477168 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:35Z","lastTransitionTime":"2025-12-01T08:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.580090 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.580154 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.580174 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.580199 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.580218 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:35Z","lastTransitionTime":"2025-12-01T08:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.683557 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.684085 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.684183 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.684282 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.684380 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:35Z","lastTransitionTime":"2025-12-01T08:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.787250 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.787311 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.787329 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.787354 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.787372 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:35Z","lastTransitionTime":"2025-12-01T08:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.890514 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.890574 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.890592 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.890619 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.890639 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:35Z","lastTransitionTime":"2025-12-01T08:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.975498 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dxlvk_a308045b-f54d-42ec-8fdf-5bc3e54ef363/ovnkube-controller/3.log" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.979656 4873 scope.go:117] "RemoveContainer" containerID="c55ab10131e65cc646cead8704fd95f3534c6726b01283d2e3fdf4b8da4e03ff" Dec 01 08:41:35 crc kubenswrapper[4873]: E1201 08:41:35.979912 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dxlvk_openshift-ovn-kubernetes(a308045b-f54d-42ec-8fdf-5bc3e54ef363)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.993098 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.993138 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.993148 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.993162 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.993172 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:35Z","lastTransitionTime":"2025-12-01T08:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:35 crc kubenswrapper[4873]: I1201 08:41:35.997661 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f7ac0b4-d048-46e4-bcee-4a271503e46d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8208cd5fda4df2e3b5a8bd90a18c8aca62b16cc2c60e06f52411f24f3d60dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba3cfd8a611c687414d7ecd6b02b423b351d7ac2a4d041085d4ca3bb2b0709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ae6257c73f3632c499b5f107178f64a9cac2f4325c57b41aa75d6d4ea14dd12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5088b5bf7bc847f59ca9efbf400855a70b99f1bbbf207eb93576c3b3eabae5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:35Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.013846 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:36Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.026194 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36539974b64191f4ca087bacbd02333f90c89fc8273ab7285e9a1d840155707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:36Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.038893 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef7b114-0e07-402d-a37b-315c36011f4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26e4a6e0a8877a213e2e974d4f7a3ad84938ae1365fa45ca7621bc050470e7e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67d090dbd4e672a6d3e777da244480eaf270518c68f8eaf2292ef7c366b8da4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-scwpp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:36Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.047975 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfqcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06182c66-4be6-4e00-9a17-4917badfc997\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa8ef5ef07129ec4a7b4cf4738a01cd6444e9bf548134a9da493c7149301387\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfqcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:36Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.082565 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a2348f-009e-4cc1-b591-e1c14ff0faf2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff98ccdc9925e44e62f0d6e826ea7926dfa2a7d83d33e475a092f128b7fe920a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4676d506268107b87ebe2e21ba6e65d959a23a4d6f38bf0de588bd9c5d15e207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4751109230b67ad8948a2de4794571581d662bfd8abc6dc5c198dbc411172860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3acd19dc33ad40d8857d28bc0ccee4c99a6927e92e5029873234ecb5dc6ed18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7242d1034bd1938cb0bd750cab74db52d454ddd8af75f1e9e0a3c007dcdddd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:36Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.096006 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.096076 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.096092 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.096112 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.096126 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:36Z","lastTransitionTime":"2025-12-01T08:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.098252 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88068a1cf89e2c507b9d93d305db6af5af196bd9eec805f42d91b6fe090aba35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:36Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.111646 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:36Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.126728 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:36Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.148328 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a308045b-f54d-42ec-8fdf-5bc3e54ef363\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55ab10131e65cc646cead8704fd95f3534c6726b01283d2e3fdf4b8da4e03ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c55ab10131e65cc646cead8704fd95f3534c6726b01283d2e3fdf4b8da4e03ff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:41:34Z\\\",\\\"message\\\":\\\"Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5smrf after 0 failed attempt(s)\\\\nI1201 08:41:34.466741 6856 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5smrf\\\\nI1201 08:41:34.466681 6856 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-shkn9\\\\nI1201 08:41:34.466583 6856 default_network_controller.go:776] Recording success event on pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1201 08:41:34.466600 6856 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nI1201 08:41:34.466782 6856 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-qpr6r in node crc\\\\nI1201 08:41:34.466816 6856 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI1201 08:41:34.466591 6856 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nF1201 08:41:34.466439 6856 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:41:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dxlvk_openshift-ovn-kubernetes(a308045b-f54d-42ec-8fdf-5bc3e54ef363)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dxlvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:36Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.166116 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5smrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdde424-6124-432b-9902-cedb9dca26c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6c098fa4f45096e671331702f2995119b74cc8c87c5fa1b175099070bc9099\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lvr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cfdfa53bea0d7e308f7673baf05e104ad7967e0473e36a74a17825433a17c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lvr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5smrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:36Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.180390 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9vq2r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08d02e64-a000-4f11-837e-82ad06d925f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzwvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzwvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9vq2r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:36Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.194448 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6938caf5-0969-4f3a-ba82-ae50109d7be0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25326d274f24d0da05778ed8c5a80d465065df87cf625be2c35d9c95763d4aab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9ffbc1a66dd17bac0e36d480467c72521d7a97c5376dafe57fa15efe571c78b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ce4df67c597a961222280dff543040c444f4d7eace37c5cfe07c7f9d89ea70a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aae94f574fbe3c1c7fcd3dbe889f7aa484e1482cee011ebfb2044151d4ca8eac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aae94f574fbe3c1c7fcd3dbe889f7aa484e1482cee011ebfb2044151d4ca8eac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:36Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.198913 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.198972 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.198990 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.199047 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.199068 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:36Z","lastTransitionTime":"2025-12-01T08:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.214727 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1c8f149-f188-4759-b2dc-9dc15be1bb13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:40:37.024907 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:40:37.026665 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1609525021/tls.crt::/tmp/serving-cert-1609525021/tls.key\\\\\\\"\\\\nI1201 08:40:42.169070 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:40:42.177896 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:40:42.177940 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:40:42.177968 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:40:42.177975 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:40:42.184767 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1201 08:40:42.184781 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1201 08:40:42.184802 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184808 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184814 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:40:42.184819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:40:42.184822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:40:42.184826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1201 08:40:42.187129 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:36Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.237041 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tpwwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13a2246b-93bb-4586-98ee-53fc84aaae02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9f46a2651e067cd11996fb1c43221a21ab71fdc3ca78a61d9245516f871a5f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dfcc763fe6ab69cab09f09c92ad9a85a4ded49c1443904134f78d1ca0092dae\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:41:30Z\\\",\\\"message\\\":\\\"2025-12-01T08:40:45+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_81306c4a-540c-48bc-b1a5-86f626f67123\\\\n2025-12-01T08:40:45+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_81306c4a-540c-48bc-b1a5-86f626f67123 to /host/opt/cni/bin/\\\\n2025-12-01T08:40:45Z [verbose] multus-daemon started\\\\n2025-12-01T08:40:45Z [verbose] Readiness Indicator file check\\\\n2025-12-01T08:41:30Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zfr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tpwwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:36Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.258518 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90f08b6fa658ce25fbc25d8d44d956cdb6424a0b047205f19d1a6fd6870b1e90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8d27e3678f1f63f55bd1b76668e328f49368d67d4d7338a6326cb73d13e08f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:36Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.270285 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qpr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c42ab88e96b59d9430a69343e06a329447ae0163cb0f5b9ef2bcafa1bf356583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qpr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:36Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.289265 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8570b876-a13b-4b8f-99c4-4900cfe19a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0afb7d55f395ee39aa245516631b88aab92399973a628f91172c9e410dcc13dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shkn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:36Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.302234 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.302279 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.302293 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.302314 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.302329 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:36Z","lastTransitionTime":"2025-12-01T08:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.404971 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.405031 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.405052 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.405069 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.405080 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:36Z","lastTransitionTime":"2025-12-01T08:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.429455 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.429492 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.429508 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.429611 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:36 crc kubenswrapper[4873]: E1201 08:41:36.429632 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:36 crc kubenswrapper[4873]: E1201 08:41:36.429789 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:41:36 crc kubenswrapper[4873]: E1201 08:41:36.429888 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:36 crc kubenswrapper[4873]: E1201 08:41:36.429953 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.507841 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.507907 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.507921 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.507945 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.507959 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:36Z","lastTransitionTime":"2025-12-01T08:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.610581 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.610628 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.610637 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.610655 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.610667 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:36Z","lastTransitionTime":"2025-12-01T08:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.713925 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.713983 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.713997 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.714036 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.714051 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:36Z","lastTransitionTime":"2025-12-01T08:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.715154 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.715187 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.715199 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.715214 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.715227 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:36Z","lastTransitionTime":"2025-12-01T08:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:36 crc kubenswrapper[4873]: E1201 08:41:36.731070 4873 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963d6124-09f6-48b8-b38c-854877aa92e3\\\",\\\"systemUUID\\\":\\\"27c66b0b-2450-400c-b6ea-cdb9cbbf95ad\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:36Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.735788 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.735839 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.735860 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.735884 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.735923 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:36Z","lastTransitionTime":"2025-12-01T08:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:36 crc kubenswrapper[4873]: E1201 08:41:36.751582 4873 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963d6124-09f6-48b8-b38c-854877aa92e3\\\",\\\"systemUUID\\\":\\\"27c66b0b-2450-400c-b6ea-cdb9cbbf95ad\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:36Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.756049 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.756105 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.756124 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.756145 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.756161 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:36Z","lastTransitionTime":"2025-12-01T08:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:36 crc kubenswrapper[4873]: E1201 08:41:36.771599 4873 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963d6124-09f6-48b8-b38c-854877aa92e3\\\",\\\"systemUUID\\\":\\\"27c66b0b-2450-400c-b6ea-cdb9cbbf95ad\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:36Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.776697 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.776753 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.776765 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.776785 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.776798 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:36Z","lastTransitionTime":"2025-12-01T08:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:36 crc kubenswrapper[4873]: E1201 08:41:36.791627 4873 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963d6124-09f6-48b8-b38c-854877aa92e3\\\",\\\"systemUUID\\\":\\\"27c66b0b-2450-400c-b6ea-cdb9cbbf95ad\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:36Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.797196 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.797249 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.797266 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.797291 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.797309 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:36Z","lastTransitionTime":"2025-12-01T08:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:36 crc kubenswrapper[4873]: E1201 08:41:36.823436 4873 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963d6124-09f6-48b8-b38c-854877aa92e3\\\",\\\"systemUUID\\\":\\\"27c66b0b-2450-400c-b6ea-cdb9cbbf95ad\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:36Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:36 crc kubenswrapper[4873]: E1201 08:41:36.823578 4873 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.825768 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.825883 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.825907 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.825936 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.825961 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:36Z","lastTransitionTime":"2025-12-01T08:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.929247 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.929321 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.929343 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.929372 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:36 crc kubenswrapper[4873]: I1201 08:41:36.929392 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:36Z","lastTransitionTime":"2025-12-01T08:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.032151 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.032186 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.032195 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.032211 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.032219 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:37Z","lastTransitionTime":"2025-12-01T08:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.135100 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.135157 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.135168 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.135189 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.135203 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:37Z","lastTransitionTime":"2025-12-01T08:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.238590 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.238677 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.238697 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.238751 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.238772 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:37Z","lastTransitionTime":"2025-12-01T08:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.342679 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.342764 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.342785 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.342816 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.342835 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:37Z","lastTransitionTime":"2025-12-01T08:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.447043 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.447492 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.447644 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.447862 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.448020 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:37Z","lastTransitionTime":"2025-12-01T08:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.551653 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.551723 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.551742 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.551767 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.551787 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:37Z","lastTransitionTime":"2025-12-01T08:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.655893 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.655953 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.655971 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.655995 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.656015 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:37Z","lastTransitionTime":"2025-12-01T08:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.759671 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.759731 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.759750 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.759774 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.759794 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:37Z","lastTransitionTime":"2025-12-01T08:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.863641 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.863765 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.863788 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.863820 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.863857 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:37Z","lastTransitionTime":"2025-12-01T08:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.967369 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.967407 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.967415 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.967429 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:37 crc kubenswrapper[4873]: I1201 08:41:37.967438 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:37Z","lastTransitionTime":"2025-12-01T08:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.072134 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.072177 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.072190 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.072213 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.072227 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:38Z","lastTransitionTime":"2025-12-01T08:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.176102 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.176169 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.176209 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.176242 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.176265 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:38Z","lastTransitionTime":"2025-12-01T08:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.280116 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.280198 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.280219 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.280245 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.280265 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:38Z","lastTransitionTime":"2025-12-01T08:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.384957 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.385074 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.385099 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.385136 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.385161 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:38Z","lastTransitionTime":"2025-12-01T08:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.429929 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.430136 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.430135 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:38 crc kubenswrapper[4873]: E1201 08:41:38.430348 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.430380 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:38 crc kubenswrapper[4873]: E1201 08:41:38.430505 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:41:38 crc kubenswrapper[4873]: E1201 08:41:38.430632 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:38 crc kubenswrapper[4873]: E1201 08:41:38.430723 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.488980 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.489055 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.489071 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.489092 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.489107 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:38Z","lastTransitionTime":"2025-12-01T08:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.591965 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.592045 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.592062 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.592085 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.592101 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:38Z","lastTransitionTime":"2025-12-01T08:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.695780 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.695844 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.695863 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.695888 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.695905 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:38Z","lastTransitionTime":"2025-12-01T08:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.800137 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.800209 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.800230 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.800257 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.800280 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:38Z","lastTransitionTime":"2025-12-01T08:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.903559 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.903622 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.903645 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.903675 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:38 crc kubenswrapper[4873]: I1201 08:41:38.903697 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:38Z","lastTransitionTime":"2025-12-01T08:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.006157 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.006221 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.006240 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.006267 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.006285 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:39Z","lastTransitionTime":"2025-12-01T08:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.109924 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.109996 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.110024 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.110042 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.110057 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:39Z","lastTransitionTime":"2025-12-01T08:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.213646 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.213694 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.213705 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.213723 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.213736 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:39Z","lastTransitionTime":"2025-12-01T08:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.316470 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.316548 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.316570 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.316593 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.316613 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:39Z","lastTransitionTime":"2025-12-01T08:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.420560 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.420645 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.420671 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.420707 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.420734 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:39Z","lastTransitionTime":"2025-12-01T08:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.447341 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.523587 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.523732 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.523805 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.523833 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.523893 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:39Z","lastTransitionTime":"2025-12-01T08:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.627253 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.627315 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.627332 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.627358 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.627379 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:39Z","lastTransitionTime":"2025-12-01T08:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.731710 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.731789 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.731813 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.731843 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.731864 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:39Z","lastTransitionTime":"2025-12-01T08:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.834957 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.835077 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.835101 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.835130 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.835149 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:39Z","lastTransitionTime":"2025-12-01T08:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.938454 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.938507 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.938523 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.938548 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:39 crc kubenswrapper[4873]: I1201 08:41:39.938566 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:39Z","lastTransitionTime":"2025-12-01T08:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.041635 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.041691 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.041709 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.041730 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.041747 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:40Z","lastTransitionTime":"2025-12-01T08:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.146335 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.146435 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.146457 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.146480 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.146507 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:40Z","lastTransitionTime":"2025-12-01T08:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.249223 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.249289 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.249343 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.249368 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.249385 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:40Z","lastTransitionTime":"2025-12-01T08:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.352368 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.352452 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.352476 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.352504 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.352526 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:40Z","lastTransitionTime":"2025-12-01T08:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.455995 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.456147 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.456169 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.456196 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.456215 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:40Z","lastTransitionTime":"2025-12-01T08:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.478384 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:41:40 crc kubenswrapper[4873]: E1201 08:41:40.479694 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.479779 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.479801 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:40 crc kubenswrapper[4873]: E1201 08:41:40.479945 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.479962 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:40 crc kubenswrapper[4873]: E1201 08:41:40.480157 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:40 crc kubenswrapper[4873]: E1201 08:41:40.480302 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.559396 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.559463 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.559475 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.559497 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.559511 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:40Z","lastTransitionTime":"2025-12-01T08:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.664098 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.664178 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.664198 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.664223 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.664241 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:40Z","lastTransitionTime":"2025-12-01T08:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.767798 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.767855 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.767872 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.767898 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.767915 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:40Z","lastTransitionTime":"2025-12-01T08:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.871309 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.871432 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.871456 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.871482 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.871504 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:40Z","lastTransitionTime":"2025-12-01T08:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.974663 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.974750 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.974777 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.974808 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:40 crc kubenswrapper[4873]: I1201 08:41:40.974829 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:40Z","lastTransitionTime":"2025-12-01T08:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:41 crc kubenswrapper[4873]: I1201 08:41:41.077868 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:41 crc kubenswrapper[4873]: I1201 08:41:41.077987 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:41 crc kubenswrapper[4873]: I1201 08:41:41.078016 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:41 crc kubenswrapper[4873]: I1201 08:41:41.078102 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:41 crc kubenswrapper[4873]: I1201 08:41:41.078124 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:41Z","lastTransitionTime":"2025-12-01T08:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:41 crc kubenswrapper[4873]: I1201 08:41:41.182164 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:41 crc kubenswrapper[4873]: I1201 08:41:41.182223 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:41 crc kubenswrapper[4873]: I1201 08:41:41.182245 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:41 crc kubenswrapper[4873]: I1201 08:41:41.182271 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:41 crc kubenswrapper[4873]: I1201 08:41:41.182291 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:41Z","lastTransitionTime":"2025-12-01T08:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:41 crc kubenswrapper[4873]: I1201 08:41:41.285774 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:41 crc kubenswrapper[4873]: I1201 08:41:41.285829 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:41 crc kubenswrapper[4873]: I1201 08:41:41.285847 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:41 crc kubenswrapper[4873]: I1201 08:41:41.285874 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:41 crc kubenswrapper[4873]: I1201 08:41:41.285894 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:41Z","lastTransitionTime":"2025-12-01T08:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:41 crc kubenswrapper[4873]: I1201 08:41:41.389445 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:41 crc kubenswrapper[4873]: I1201 08:41:41.389653 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:41 crc kubenswrapper[4873]: I1201 08:41:41.389688 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:41 crc kubenswrapper[4873]: I1201 08:41:41.389722 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:41 crc kubenswrapper[4873]: I1201 08:41:41.389744 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:41Z","lastTransitionTime":"2025-12-01T08:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:41 crc kubenswrapper[4873]: I1201 08:41:41.493784 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:41 crc kubenswrapper[4873]: I1201 08:41:41.493849 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:41 crc kubenswrapper[4873]: I1201 08:41:41.493869 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:41 crc kubenswrapper[4873]: I1201 08:41:41.493896 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:41 crc kubenswrapper[4873]: I1201 08:41:41.493914 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:41Z","lastTransitionTime":"2025-12-01T08:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:41 crc kubenswrapper[4873]: I1201 08:41:41.596999 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:41 crc kubenswrapper[4873]: I1201 08:41:41.597118 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:41 crc kubenswrapper[4873]: I1201 08:41:41.597143 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:41 crc kubenswrapper[4873]: I1201 08:41:41.597171 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:41 crc kubenswrapper[4873]: I1201 08:41:41.597192 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:41Z","lastTransitionTime":"2025-12-01T08:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:41 crc kubenswrapper[4873]: I1201 08:41:41.700764 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:41 crc kubenswrapper[4873]: I1201 08:41:41.700829 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:41 crc kubenswrapper[4873]: I1201 08:41:41.700848 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:41 crc kubenswrapper[4873]: I1201 08:41:41.700873 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:41 crc kubenswrapper[4873]: I1201 08:41:41.700891 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:41Z","lastTransitionTime":"2025-12-01T08:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:41 crc kubenswrapper[4873]: I1201 08:41:41.804395 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:41 crc kubenswrapper[4873]: I1201 08:41:41.804448 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:41 crc kubenswrapper[4873]: I1201 08:41:41.804461 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:41 crc kubenswrapper[4873]: I1201 08:41:41.804480 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:41 crc kubenswrapper[4873]: I1201 08:41:41.804494 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:41Z","lastTransitionTime":"2025-12-01T08:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:41 crc kubenswrapper[4873]: I1201 08:41:41.907034 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:41 crc kubenswrapper[4873]: I1201 08:41:41.907116 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:41 crc kubenswrapper[4873]: I1201 08:41:41.907128 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:41 crc kubenswrapper[4873]: I1201 08:41:41.907150 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:41 crc kubenswrapper[4873]: I1201 08:41:41.907163 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:41Z","lastTransitionTime":"2025-12-01T08:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.012447 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.012525 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.012542 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.012567 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.012586 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:42Z","lastTransitionTime":"2025-12-01T08:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.115639 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.115683 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.115694 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.115710 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.115721 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:42Z","lastTransitionTime":"2025-12-01T08:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.219490 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.219582 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.219598 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.219622 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.219637 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:42Z","lastTransitionTime":"2025-12-01T08:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.323259 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.323341 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.323366 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.323396 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.323419 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:42Z","lastTransitionTime":"2025-12-01T08:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.426437 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.426514 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.426537 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.426565 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.426586 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:42Z","lastTransitionTime":"2025-12-01T08:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.429887 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.429977 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.429984 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.430142 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:42 crc kubenswrapper[4873]: E1201 08:41:42.430152 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:42 crc kubenswrapper[4873]: E1201 08:41:42.430314 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:41:42 crc kubenswrapper[4873]: E1201 08:41:42.430450 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:42 crc kubenswrapper[4873]: E1201 08:41:42.430624 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.528437 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.528478 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.528488 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.528502 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.528512 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:42Z","lastTransitionTime":"2025-12-01T08:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.632187 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.632252 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.632273 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.632304 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.632331 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:42Z","lastTransitionTime":"2025-12-01T08:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.735193 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.735259 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.735280 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.735310 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.735331 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:42Z","lastTransitionTime":"2025-12-01T08:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.838916 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.838976 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.838997 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.839088 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.839124 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:42Z","lastTransitionTime":"2025-12-01T08:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.942903 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.942978 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.943001 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.943067 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:42 crc kubenswrapper[4873]: I1201 08:41:42.943088 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:42Z","lastTransitionTime":"2025-12-01T08:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.046130 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.046196 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.046213 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.046237 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.046250 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:43Z","lastTransitionTime":"2025-12-01T08:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.149966 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.150086 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.150109 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.150138 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.150156 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:43Z","lastTransitionTime":"2025-12-01T08:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.254375 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.254453 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.254471 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.254497 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.254517 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:43Z","lastTransitionTime":"2025-12-01T08:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.357674 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.357840 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.357872 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.357906 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.357931 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:43Z","lastTransitionTime":"2025-12-01T08:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.461337 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.461423 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.461449 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.461481 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.461501 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:43Z","lastTransitionTime":"2025-12-01T08:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.563918 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.563981 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.563993 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.564038 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.564054 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:43Z","lastTransitionTime":"2025-12-01T08:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.666912 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.666988 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.667028 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.667051 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.667065 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:43Z","lastTransitionTime":"2025-12-01T08:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.770930 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.770987 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.771004 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.771046 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.771062 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:43Z","lastTransitionTime":"2025-12-01T08:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.873623 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.873717 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.873738 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.873765 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.873783 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:43Z","lastTransitionTime":"2025-12-01T08:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.976749 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.976830 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.976845 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.976864 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:43 crc kubenswrapper[4873]: I1201 08:41:43.976878 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:43Z","lastTransitionTime":"2025-12-01T08:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.079957 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.080032 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.080043 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.080064 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.080078 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:44Z","lastTransitionTime":"2025-12-01T08:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.184092 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.184174 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.184197 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.184225 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.184256 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:44Z","lastTransitionTime":"2025-12-01T08:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.288689 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.288793 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.288847 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.288950 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.288984 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:44Z","lastTransitionTime":"2025-12-01T08:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.391282 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.391379 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.391397 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.391426 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.391446 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:44Z","lastTransitionTime":"2025-12-01T08:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.429644 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.429692 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:44 crc kubenswrapper[4873]: E1201 08:41:44.429749 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.429923 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:41:44 crc kubenswrapper[4873]: E1201 08:41:44.430077 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.430093 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:44 crc kubenswrapper[4873]: E1201 08:41:44.430259 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:41:44 crc kubenswrapper[4873]: E1201 08:41:44.430311 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.466618 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a2348f-009e-4cc1-b591-e1c14ff0faf2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff98ccdc9925e44e62f0d6e826ea7926dfa2a7d83d33e475a092f128b7fe920a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4676d506268107b87ebe2e21ba6e65d959a23a4d6f38bf0de588bd9c5d15e207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4751109230b67ad8948a2de4794571581d662bfd8abc6dc5c198dbc411172860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3acd19dc33ad40d8857d28bc0ccee4c99a6927e92e5029873234ecb5dc6ed18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7242d1034bd1938cb0bd750cab74db52d454ddd8af75f1e9e0a3c007dcdddd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.488512 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88068a1cf89e2c507b9d93d305db6af5af196bd9eec805f42d91b6fe090aba35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.494356 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.494423 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.494449 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.494481 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.494505 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:44Z","lastTransitionTime":"2025-12-01T08:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.508282 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.524750 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.558285 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a308045b-f54d-42ec-8fdf-5bc3e54ef363\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55ab10131e65cc646cead8704fd95f3534c6726b01283d2e3fdf4b8da4e03ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c55ab10131e65cc646cead8704fd95f3534c6726b01283d2e3fdf4b8da4e03ff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:41:34Z\\\",\\\"message\\\":\\\"Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5smrf after 0 failed attempt(s)\\\\nI1201 08:41:34.466741 6856 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5smrf\\\\nI1201 08:41:34.466681 6856 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-shkn9\\\\nI1201 08:41:34.466583 6856 default_network_controller.go:776] Recording success event on pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1201 08:41:34.466600 6856 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nI1201 08:41:34.466782 6856 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-qpr6r in node crc\\\\nI1201 08:41:34.466816 6856 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI1201 08:41:34.466591 6856 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nF1201 08:41:34.466439 6856 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:41:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dxlvk_openshift-ovn-kubernetes(a308045b-f54d-42ec-8fdf-5bc3e54ef363)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dxlvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.575544 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5smrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdde424-6124-432b-9902-cedb9dca26c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6c098fa4f45096e671331702f2995119b74cc8c87c5fa1b175099070bc9099\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lvr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cfdfa53bea0d7e308f7673baf05e104ad7967e0473e36a74a17825433a17c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lvr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5smrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.591215 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9vq2r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08d02e64-a000-4f11-837e-82ad06d925f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzwvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzwvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9vq2r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.597529 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.597616 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.597647 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.597676 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.597696 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:44Z","lastTransitionTime":"2025-12-01T08:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.612458 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6938caf5-0969-4f3a-ba82-ae50109d7be0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25326d274f24d0da05778ed8c5a80d465065df87cf625be2c35d9c95763d4aab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9ffbc1a66dd17bac0e36d480467c72521d7a97c5376dafe57fa15efe571c78b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ce4df67c597a961222280dff543040c444f4d7eace37c5cfe07c7f9d89ea70a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aae94f574fbe3c1c7fcd3dbe889f7aa484e1482cee011ebfb2044151d4ca8eac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aae94f574fbe3c1c7fcd3dbe889f7aa484e1482cee011ebfb2044151d4ca8eac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.627869 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fc7f72c-f0d9-42ac-80a3-770de844f482\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd0a7d74f98d7769eb5513cf477937a5c5f23bc6374335d17281087987d332c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24642ae22d2e061a961799d5ca7e07e31897891e2a43cf671bea739030079fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24642ae22d2e061a961799d5ca7e07e31897891e2a43cf671bea739030079fab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.646120 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1c8f149-f188-4759-b2dc-9dc15be1bb13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:40:37.024907 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:40:37.026665 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1609525021/tls.crt::/tmp/serving-cert-1609525021/tls.key\\\\\\\"\\\\nI1201 08:40:42.169070 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:40:42.177896 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:40:42.177940 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:40:42.177968 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:40:42.177975 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:40:42.184767 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1201 08:40:42.184781 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1201 08:40:42.184802 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184808 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184814 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:40:42.184819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:40:42.184822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:40:42.184826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1201 08:40:42.187129 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.666459 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tpwwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13a2246b-93bb-4586-98ee-53fc84aaae02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9f46a2651e067cd11996fb1c43221a21ab71fdc3ca78a61d9245516f871a5f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dfcc763fe6ab69cab09f09c92ad9a85a4ded49c1443904134f78d1ca0092dae\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:41:30Z\\\",\\\"message\\\":\\\"2025-12-01T08:40:45+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_81306c4a-540c-48bc-b1a5-86f626f67123\\\\n2025-12-01T08:40:45+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_81306c4a-540c-48bc-b1a5-86f626f67123 to /host/opt/cni/bin/\\\\n2025-12-01T08:40:45Z [verbose] multus-daemon started\\\\n2025-12-01T08:40:45Z [verbose] Readiness Indicator file check\\\\n2025-12-01T08:41:30Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zfr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tpwwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.681790 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90f08b6fa658ce25fbc25d8d44d956cdb6424a0b047205f19d1a6fd6870b1e90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8d27e3678f1f63f55bd1b76668e328f49368d67d4d7338a6326cb73d13e08f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.699554 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qpr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c42ab88e96b59d9430a69343e06a329447ae0163cb0f5b9ef2bcafa1bf356583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qpr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.702457 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.702587 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.702610 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.702656 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.702674 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:44Z","lastTransitionTime":"2025-12-01T08:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.717216 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8570b876-a13b-4b8f-99c4-4900cfe19a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0afb7d55f395ee39aa245516631b88aab92399973a628f91172c9e410dcc13dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shkn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.734999 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f7ac0b4-d048-46e4-bcee-4a271503e46d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8208cd5fda4df2e3b5a8bd90a18c8aca62b16cc2c60e06f52411f24f3d60dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba3cfd8a611c687414d7ecd6b02b423b351d7ac2a4d041085d4ca3bb2b0709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ae6257c73f3632c499b5f107178f64a9cac2f4325c57b41aa75d6d4ea14dd12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5088b5bf7bc847f59ca9efbf400855a70b99f1bbbf207eb93576c3b3eabae5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.751680 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.763924 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36539974b64191f4ca087bacbd02333f90c89fc8273ab7285e9a1d840155707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.780378 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef7b114-0e07-402d-a37b-315c36011f4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26e4a6e0a8877a213e2e974d4f7a3ad84938ae1365fa45ca7621bc050470e7e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67d090dbd4e672a6d3e777da244480eaf270518c68f8eaf2292ef7c366b8da4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-scwpp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.793795 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfqcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06182c66-4be6-4e00-9a17-4917badfc997\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa8ef5ef07129ec4a7b4cf4738a01cd6444e9bf548134a9da493c7149301387\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfqcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.806332 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.806411 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.806427 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.806447 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.806463 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:44Z","lastTransitionTime":"2025-12-01T08:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.909505 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.909581 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.909605 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.909634 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:44 crc kubenswrapper[4873]: I1201 08:41:44.909654 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:44Z","lastTransitionTime":"2025-12-01T08:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.012310 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.012366 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.012381 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.012403 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.012415 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:45Z","lastTransitionTime":"2025-12-01T08:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.115526 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.115595 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.115613 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.115641 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.115659 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:45Z","lastTransitionTime":"2025-12-01T08:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.220201 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.220303 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.220332 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.220368 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.220393 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:45Z","lastTransitionTime":"2025-12-01T08:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.323629 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.323711 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.323735 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.323763 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.323781 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:45Z","lastTransitionTime":"2025-12-01T08:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.426757 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.426835 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.426859 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.426888 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.426912 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:45Z","lastTransitionTime":"2025-12-01T08:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.530355 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.530399 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.530410 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.530427 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.530438 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:45Z","lastTransitionTime":"2025-12-01T08:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.633810 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.633890 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.633914 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.633945 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.633968 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:45Z","lastTransitionTime":"2025-12-01T08:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.737772 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.737839 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.737863 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.737890 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.737910 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:45Z","lastTransitionTime":"2025-12-01T08:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.841816 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.841883 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.841906 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.841935 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.841957 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:45Z","lastTransitionTime":"2025-12-01T08:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.945682 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.945741 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.945758 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.945783 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:45 crc kubenswrapper[4873]: I1201 08:41:45.945835 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:45Z","lastTransitionTime":"2025-12-01T08:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.048554 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.048650 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.048676 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.048707 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.048729 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:46Z","lastTransitionTime":"2025-12-01T08:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.153137 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.153828 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.153850 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.153879 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.153901 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:46Z","lastTransitionTime":"2025-12-01T08:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.256474 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.256548 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.256569 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.256597 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.256621 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:46Z","lastTransitionTime":"2025-12-01T08:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.360464 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.360535 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.360553 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.360578 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.360597 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:46Z","lastTransitionTime":"2025-12-01T08:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.395246 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.395552 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.395619 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:46 crc kubenswrapper[4873]: E1201 08:41:46.395666 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:50.395622767 +0000 UTC m=+146.297731346 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:46 crc kubenswrapper[4873]: E1201 08:41:46.395719 4873 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.395783 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:46 crc kubenswrapper[4873]: E1201 08:41:46.395823 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:42:50.395787461 +0000 UTC m=+146.297896040 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:41:46 crc kubenswrapper[4873]: E1201 08:41:46.395819 4873 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.395870 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:46 crc kubenswrapper[4873]: E1201 08:41:46.395920 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:42:50.395891844 +0000 UTC m=+146.298000423 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:41:46 crc kubenswrapper[4873]: E1201 08:41:46.395969 4873 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:41:46 crc kubenswrapper[4873]: E1201 08:41:46.395995 4873 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:41:46 crc kubenswrapper[4873]: E1201 08:41:46.396059 4873 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:41:46 crc kubenswrapper[4873]: E1201 08:41:46.396087 4873 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:41:46 crc kubenswrapper[4873]: E1201 08:41:46.396110 4873 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:41:46 crc kubenswrapper[4873]: E1201 08:41:46.396132 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 08:42:50.3961134 +0000 UTC m=+146.298221979 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:41:46 crc kubenswrapper[4873]: E1201 08:41:46.396133 4873 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:41:46 crc kubenswrapper[4873]: E1201 08:41:46.396189 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 08:42:50.396175421 +0000 UTC m=+146.298283990 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.430006 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.430091 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.430006 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.430213 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:46 crc kubenswrapper[4873]: E1201 08:41:46.430378 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:46 crc kubenswrapper[4873]: E1201 08:41:46.430577 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:41:46 crc kubenswrapper[4873]: E1201 08:41:46.430807 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:46 crc kubenswrapper[4873]: E1201 08:41:46.430909 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.468533 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.468626 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.468687 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.468746 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.468771 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:46Z","lastTransitionTime":"2025-12-01T08:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.572275 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.572359 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.572385 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.572418 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.572442 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:46Z","lastTransitionTime":"2025-12-01T08:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.676799 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.676877 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.676903 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.676936 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.676958 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:46Z","lastTransitionTime":"2025-12-01T08:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.781223 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.781345 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.781369 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.781393 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.781412 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:46Z","lastTransitionTime":"2025-12-01T08:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.884651 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.884780 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.884801 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.884824 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.884841 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:46Z","lastTransitionTime":"2025-12-01T08:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.988720 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.988781 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.988799 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.988824 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:46 crc kubenswrapper[4873]: I1201 08:41:46.988844 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:46Z","lastTransitionTime":"2025-12-01T08:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.011844 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.011883 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.011899 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.011921 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.011937 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:47Z","lastTransitionTime":"2025-12-01T08:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:47 crc kubenswrapper[4873]: E1201 08:41:47.033800 4873 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963d6124-09f6-48b8-b38c-854877aa92e3\\\",\\\"systemUUID\\\":\\\"27c66b0b-2450-400c-b6ea-cdb9cbbf95ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:47Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.039430 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.039490 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.039513 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.039547 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.039569 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:47Z","lastTransitionTime":"2025-12-01T08:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:47 crc kubenswrapper[4873]: E1201 08:41:47.062450 4873 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963d6124-09f6-48b8-b38c-854877aa92e3\\\",\\\"systemUUID\\\":\\\"27c66b0b-2450-400c-b6ea-cdb9cbbf95ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:47Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.068742 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.068817 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.068841 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.068872 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.068898 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:47Z","lastTransitionTime":"2025-12-01T08:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:47 crc kubenswrapper[4873]: E1201 08:41:47.090087 4873 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963d6124-09f6-48b8-b38c-854877aa92e3\\\",\\\"systemUUID\\\":\\\"27c66b0b-2450-400c-b6ea-cdb9cbbf95ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:47Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.096149 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.096200 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.096223 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.096251 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.096274 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:47Z","lastTransitionTime":"2025-12-01T08:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:47 crc kubenswrapper[4873]: E1201 08:41:47.117466 4873 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963d6124-09f6-48b8-b38c-854877aa92e3\\\",\\\"systemUUID\\\":\\\"27c66b0b-2450-400c-b6ea-cdb9cbbf95ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:47Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.123468 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.123563 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.123599 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.123631 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.123655 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:47Z","lastTransitionTime":"2025-12-01T08:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:47 crc kubenswrapper[4873]: E1201 08:41:47.147100 4873 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963d6124-09f6-48b8-b38c-854877aa92e3\\\",\\\"systemUUID\\\":\\\"27c66b0b-2450-400c-b6ea-cdb9cbbf95ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:47Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:47 crc kubenswrapper[4873]: E1201 08:41:47.147325 4873 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.149954 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.150013 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.150068 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.150100 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.150120 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:47Z","lastTransitionTime":"2025-12-01T08:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.253538 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.253601 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.253619 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.253643 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.253661 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:47Z","lastTransitionTime":"2025-12-01T08:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.356695 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.356762 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.356790 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.356823 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.356848 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:47Z","lastTransitionTime":"2025-12-01T08:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.460256 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.460307 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.460324 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.460344 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.460364 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:47Z","lastTransitionTime":"2025-12-01T08:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.564183 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.564261 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.564281 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.564310 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.564329 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:47Z","lastTransitionTime":"2025-12-01T08:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.667859 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.667940 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.667966 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.668001 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.668075 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:47Z","lastTransitionTime":"2025-12-01T08:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.771223 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.771289 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.771306 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.771331 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.771351 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:47Z","lastTransitionTime":"2025-12-01T08:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.873889 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.873950 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.873967 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.873989 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.874008 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:47Z","lastTransitionTime":"2025-12-01T08:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.976941 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.976990 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.977000 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.977037 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:47 crc kubenswrapper[4873]: I1201 08:41:47.977051 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:47Z","lastTransitionTime":"2025-12-01T08:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.081505 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.081620 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.081641 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.081667 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.081686 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:48Z","lastTransitionTime":"2025-12-01T08:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.185320 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.185364 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.185375 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.185394 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.185406 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:48Z","lastTransitionTime":"2025-12-01T08:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.288471 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.288535 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.288550 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.288569 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.288582 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:48Z","lastTransitionTime":"2025-12-01T08:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.391691 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.391727 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.391736 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.391750 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.391760 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:48Z","lastTransitionTime":"2025-12-01T08:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.429609 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.429628 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:48 crc kubenswrapper[4873]: E1201 08:41:48.429911 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.430191 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.431065 4873 scope.go:117] "RemoveContainer" containerID="c55ab10131e65cc646cead8704fd95f3534c6726b01283d2e3fdf4b8da4e03ff" Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.431149 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:41:48 crc kubenswrapper[4873]: E1201 08:41:48.431315 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dxlvk_openshift-ovn-kubernetes(a308045b-f54d-42ec-8fdf-5bc3e54ef363)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" Dec 01 08:41:48 crc kubenswrapper[4873]: E1201 08:41:48.431363 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:41:48 crc kubenswrapper[4873]: E1201 08:41:48.431481 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:48 crc kubenswrapper[4873]: E1201 08:41:48.431525 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.495303 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.495483 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.495598 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.495693 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.495796 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:48Z","lastTransitionTime":"2025-12-01T08:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.599192 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.599259 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.599279 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.599302 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.599321 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:48Z","lastTransitionTime":"2025-12-01T08:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.702159 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.702237 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.702257 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.702286 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.702305 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:48Z","lastTransitionTime":"2025-12-01T08:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.808715 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.808764 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.808777 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.808795 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.808809 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:48Z","lastTransitionTime":"2025-12-01T08:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.912844 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.912985 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.913084 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.913164 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:48 crc kubenswrapper[4873]: I1201 08:41:48.913191 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:48Z","lastTransitionTime":"2025-12-01T08:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.016584 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.016653 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.016669 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.016690 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.016703 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:49Z","lastTransitionTime":"2025-12-01T08:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.120341 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.120418 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.120459 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.120495 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.120520 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:49Z","lastTransitionTime":"2025-12-01T08:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.224636 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.224711 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.224730 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.224756 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.224792 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:49Z","lastTransitionTime":"2025-12-01T08:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.328237 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.328320 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.328344 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.328373 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.328395 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:49Z","lastTransitionTime":"2025-12-01T08:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.431989 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.432122 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.432146 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.432173 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.432192 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:49Z","lastTransitionTime":"2025-12-01T08:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.535800 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.535874 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.535900 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.535930 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.535956 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:49Z","lastTransitionTime":"2025-12-01T08:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.638864 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.638934 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.638958 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.638989 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.639012 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:49Z","lastTransitionTime":"2025-12-01T08:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.741797 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.741895 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.741912 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.741938 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.741957 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:49Z","lastTransitionTime":"2025-12-01T08:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.844808 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.844867 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.844882 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.844902 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.844915 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:49Z","lastTransitionTime":"2025-12-01T08:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.947223 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.947262 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.947271 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.947288 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:49 crc kubenswrapper[4873]: I1201 08:41:49.947297 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:49Z","lastTransitionTime":"2025-12-01T08:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.051064 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.051142 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.051154 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.051178 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.051192 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:50Z","lastTransitionTime":"2025-12-01T08:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.153804 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.153870 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.153884 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.153911 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.153923 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:50Z","lastTransitionTime":"2025-12-01T08:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.257239 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.257310 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.257329 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.257360 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.257384 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:50Z","lastTransitionTime":"2025-12-01T08:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.360766 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.360868 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.360891 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.360922 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.360949 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:50Z","lastTransitionTime":"2025-12-01T08:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.429889 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.429984 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.429977 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:41:50 crc kubenswrapper[4873]: E1201 08:41:50.430114 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.430202 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:50 crc kubenswrapper[4873]: E1201 08:41:50.430222 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:50 crc kubenswrapper[4873]: E1201 08:41:50.430396 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:41:50 crc kubenswrapper[4873]: E1201 08:41:50.430451 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.464432 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.464495 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.464505 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.464523 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.464535 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:50Z","lastTransitionTime":"2025-12-01T08:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.568253 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.568307 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.568319 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.568337 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.568351 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:50Z","lastTransitionTime":"2025-12-01T08:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.671924 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.671967 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.671979 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.671998 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.672010 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:50Z","lastTransitionTime":"2025-12-01T08:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.774735 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.774803 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.774820 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.774843 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.774861 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:50Z","lastTransitionTime":"2025-12-01T08:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.877932 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.877995 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.878040 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.878064 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.878085 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:50Z","lastTransitionTime":"2025-12-01T08:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.981601 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.981656 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.981674 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.981696 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:50 crc kubenswrapper[4873]: I1201 08:41:50.981714 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:50Z","lastTransitionTime":"2025-12-01T08:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:51 crc kubenswrapper[4873]: I1201 08:41:51.084274 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:51 crc kubenswrapper[4873]: I1201 08:41:51.084357 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:51 crc kubenswrapper[4873]: I1201 08:41:51.084379 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:51 crc kubenswrapper[4873]: I1201 08:41:51.084404 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:51 crc kubenswrapper[4873]: I1201 08:41:51.084423 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:51Z","lastTransitionTime":"2025-12-01T08:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:51 crc kubenswrapper[4873]: I1201 08:41:51.187750 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:51 crc kubenswrapper[4873]: I1201 08:41:51.187799 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:51 crc kubenswrapper[4873]: I1201 08:41:51.187815 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:51 crc kubenswrapper[4873]: I1201 08:41:51.187837 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:51 crc kubenswrapper[4873]: I1201 08:41:51.187854 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:51Z","lastTransitionTime":"2025-12-01T08:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:51 crc kubenswrapper[4873]: I1201 08:41:51.291717 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:51 crc kubenswrapper[4873]: I1201 08:41:51.291790 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:51 crc kubenswrapper[4873]: I1201 08:41:51.291808 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:51 crc kubenswrapper[4873]: I1201 08:41:51.291832 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:51 crc kubenswrapper[4873]: I1201 08:41:51.291852 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:51Z","lastTransitionTime":"2025-12-01T08:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:51 crc kubenswrapper[4873]: I1201 08:41:51.394867 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:51 crc kubenswrapper[4873]: I1201 08:41:51.394926 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:51 crc kubenswrapper[4873]: I1201 08:41:51.394944 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:51 crc kubenswrapper[4873]: I1201 08:41:51.394965 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:51 crc kubenswrapper[4873]: I1201 08:41:51.394984 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:51Z","lastTransitionTime":"2025-12-01T08:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:51 crc kubenswrapper[4873]: I1201 08:41:51.498002 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:51 crc kubenswrapper[4873]: I1201 08:41:51.498093 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:51 crc kubenswrapper[4873]: I1201 08:41:51.498116 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:51 crc kubenswrapper[4873]: I1201 08:41:51.498140 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:51 crc kubenswrapper[4873]: I1201 08:41:51.498159 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:51Z","lastTransitionTime":"2025-12-01T08:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:51 crc kubenswrapper[4873]: I1201 08:41:51.601851 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:51 crc kubenswrapper[4873]: I1201 08:41:51.601918 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:51 crc kubenswrapper[4873]: I1201 08:41:51.601937 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:51 crc kubenswrapper[4873]: I1201 08:41:51.601962 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:51 crc kubenswrapper[4873]: I1201 08:41:51.601981 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:51Z","lastTransitionTime":"2025-12-01T08:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:51 crc kubenswrapper[4873]: I1201 08:41:51.705549 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:51 crc kubenswrapper[4873]: I1201 08:41:51.705613 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:51 crc kubenswrapper[4873]: I1201 08:41:51.705636 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:51 crc kubenswrapper[4873]: I1201 08:41:51.705666 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:51 crc kubenswrapper[4873]: I1201 08:41:51.705802 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:51Z","lastTransitionTime":"2025-12-01T08:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:51 crc kubenswrapper[4873]: I1201 08:41:51.809474 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:51 crc kubenswrapper[4873]: I1201 08:41:51.809520 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:51 crc kubenswrapper[4873]: I1201 08:41:51.809536 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:51 crc kubenswrapper[4873]: I1201 08:41:51.809598 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:51 crc kubenswrapper[4873]: I1201 08:41:51.809620 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:51Z","lastTransitionTime":"2025-12-01T08:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:51 crc kubenswrapper[4873]: I1201 08:41:51.912891 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:51 crc kubenswrapper[4873]: I1201 08:41:51.912962 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:51 crc kubenswrapper[4873]: I1201 08:41:51.912979 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:51 crc kubenswrapper[4873]: I1201 08:41:51.913003 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:51 crc kubenswrapper[4873]: I1201 08:41:51.913061 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:51Z","lastTransitionTime":"2025-12-01T08:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.016590 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.016696 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.016713 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.016738 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.016758 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:52Z","lastTransitionTime":"2025-12-01T08:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.119864 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.119947 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.119971 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.120001 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.120126 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:52Z","lastTransitionTime":"2025-12-01T08:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.223322 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.223398 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.223475 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.223546 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.223578 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:52Z","lastTransitionTime":"2025-12-01T08:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.327297 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.327389 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.327406 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.327432 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.327455 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:52Z","lastTransitionTime":"2025-12-01T08:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.428989 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.429064 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.429122 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:52 crc kubenswrapper[4873]: E1201 08:41:52.429245 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.428989 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:41:52 crc kubenswrapper[4873]: E1201 08:41:52.429497 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:52 crc kubenswrapper[4873]: E1201 08:41:52.429566 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:52 crc kubenswrapper[4873]: E1201 08:41:52.429727 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.431676 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.431742 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.431765 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.431793 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.431814 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:52Z","lastTransitionTime":"2025-12-01T08:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.535964 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.536807 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.536863 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.536895 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.536909 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:52Z","lastTransitionTime":"2025-12-01T08:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.641425 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.641484 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.641504 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.641526 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.641544 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:52Z","lastTransitionTime":"2025-12-01T08:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.744741 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.744797 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.744813 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.744837 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.744854 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:52Z","lastTransitionTime":"2025-12-01T08:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.848068 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.848162 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.848182 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.848207 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.848223 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:52Z","lastTransitionTime":"2025-12-01T08:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.951836 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.951877 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.951887 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.951905 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:52 crc kubenswrapper[4873]: I1201 08:41:52.951919 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:52Z","lastTransitionTime":"2025-12-01T08:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.054791 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.054880 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.054903 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.054935 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.054962 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:53Z","lastTransitionTime":"2025-12-01T08:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.159111 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.159201 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.159226 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.159261 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.159286 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:53Z","lastTransitionTime":"2025-12-01T08:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.262900 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.262981 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.263006 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.263063 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.263082 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:53Z","lastTransitionTime":"2025-12-01T08:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.366300 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.366388 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.366422 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.366453 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.366479 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:53Z","lastTransitionTime":"2025-12-01T08:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.470374 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.470431 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.470449 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.470475 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.470509 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:53Z","lastTransitionTime":"2025-12-01T08:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.574604 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.574678 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.574696 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.574721 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.574740 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:53Z","lastTransitionTime":"2025-12-01T08:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.677374 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.677469 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.677494 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.677525 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.677549 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:53Z","lastTransitionTime":"2025-12-01T08:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.785278 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.785501 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.785581 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.785629 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.785648 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:53Z","lastTransitionTime":"2025-12-01T08:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.889358 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.889443 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.889460 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.889483 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.889498 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:53Z","lastTransitionTime":"2025-12-01T08:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.992517 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.992588 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.992603 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.992622 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:53 crc kubenswrapper[4873]: I1201 08:41:53.992635 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:53Z","lastTransitionTime":"2025-12-01T08:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.095247 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.095303 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.095326 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.095347 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.095360 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:54Z","lastTransitionTime":"2025-12-01T08:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.198979 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.199063 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.199077 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.199099 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.199112 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:54Z","lastTransitionTime":"2025-12-01T08:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.301671 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.301726 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.301745 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.301770 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.301790 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:54Z","lastTransitionTime":"2025-12-01T08:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.405618 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.405701 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.405720 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.405746 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.405767 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:54Z","lastTransitionTime":"2025-12-01T08:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.429550 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.429645 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:54 crc kubenswrapper[4873]: E1201 08:41:54.429762 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.429813 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:54 crc kubenswrapper[4873]: E1201 08:41:54.429938 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.429573 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:54 crc kubenswrapper[4873]: E1201 08:41:54.430125 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:54 crc kubenswrapper[4873]: E1201 08:41:54.430238 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.455868 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a308045b-f54d-42ec-8fdf-5bc3e54ef363\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55ab10131e65cc646cead8704fd95f3534c6726b01283d2e3fdf4b8da4e03ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c55ab10131e65cc646cead8704fd95f3534c6726b01283d2e3fdf4b8da4e03ff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:41:34Z\\\",\\\"message\\\":\\\"Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5smrf after 0 failed attempt(s)\\\\nI1201 08:41:34.466741 6856 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5smrf\\\\nI1201 08:41:34.466681 6856 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-shkn9\\\\nI1201 08:41:34.466583 6856 default_network_controller.go:776] Recording success event on pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1201 08:41:34.466600 6856 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nI1201 08:41:34.466782 6856 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-qpr6r in node crc\\\\nI1201 08:41:34.466816 6856 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI1201 08:41:34.466591 6856 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nF1201 08:41:34.466439 6856 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:41:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dxlvk_openshift-ovn-kubernetes(a308045b-f54d-42ec-8fdf-5bc3e54ef363)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxfd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dxlvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:54Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.478824 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5smrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdde424-6124-432b-9902-cedb9dca26c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6c098fa4f45096e671331702f2995119b74cc8c87c5fa1b175099070bc9099\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lvr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cfdfa53bea0d7e308f7673baf05e104ad7967e0473e36a74a17825433a17c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lvr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5smrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:54Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.495772 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9vq2r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08d02e64-a000-4f11-837e-82ad06d925f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzwvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzwvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9vq2r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:54Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.509350 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.509405 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.509424 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.509455 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.509476 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:54Z","lastTransitionTime":"2025-12-01T08:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.523031 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a2348f-009e-4cc1-b591-e1c14ff0faf2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff98ccdc9925e44e62f0d6e826ea7926dfa2a7d83d33e475a092f128b7fe920a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4676d506268107b87ebe2e21ba6e65d959a23a4d6f38bf0de588bd9c5d15e207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4751109230b67ad8948a2de4794571581d662bfd8abc6dc5c198dbc411172860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3acd19dc33ad40d8857d28bc0ccee4c99a6927e92e5029873234ecb5dc6ed18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7242d1034bd1938cb0bd750cab74db52d454ddd8af75f1e9e0a3c007dcdddd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77ade6fb2d0ef37d40cf01d0cac1cdfbbcb8432e7527f42c07c4be15da2ea268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9b38c18b030aa335363cb85cf46d6e700057d198d66c3c29bc823de0dbf0255\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3b512669a0271254403d3a10bf2be5d37f871e80e0098671be6c66b334bf1c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:54Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.543147 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88068a1cf89e2c507b9d93d305db6af5af196bd9eec805f42d91b6fe090aba35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:54Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.562339 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:54Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.578387 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:54Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.599828 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6938caf5-0969-4f3a-ba82-ae50109d7be0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25326d274f24d0da05778ed8c5a80d465065df87cf625be2c35d9c95763d4aab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9ffbc1a66dd17bac0e36d480467c72521d7a97c5376dafe57fa15efe571c78b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ce4df67c597a961222280dff543040c444f4d7eace37c5cfe07c7f9d89ea70a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aae94f574fbe3c1c7fcd3dbe889f7aa484e1482cee011ebfb2044151d4ca8eac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aae94f574fbe3c1c7fcd3dbe889f7aa484e1482cee011ebfb2044151d4ca8eac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:54Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.612925 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.612990 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.613002 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.613048 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.613065 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:54Z","lastTransitionTime":"2025-12-01T08:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.615997 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fc7f72c-f0d9-42ac-80a3-770de844f482\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd0a7d74f98d7769eb5513cf477937a5c5f23bc6374335d17281087987d332c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24642ae22d2e061a961799d5ca7e07e31897891e2a43cf671bea739030079fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24642ae22d2e061a961799d5ca7e07e31897891e2a43cf671bea739030079fab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:54Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.634734 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1c8f149-f188-4759-b2dc-9dc15be1bb13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:40:37.024907 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:40:37.026665 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1609525021/tls.crt::/tmp/serving-cert-1609525021/tls.key\\\\\\\"\\\\nI1201 08:40:42.169070 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:40:42.177896 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:40:42.177940 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:40:42.177968 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:40:42.177975 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:40:42.184767 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1201 08:40:42.184781 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1201 08:40:42.184802 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184808 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:40:42.184814 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:40:42.184819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:40:42.184822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:40:42.184826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1201 08:40:42.187129 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:54Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.656563 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tpwwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13a2246b-93bb-4586-98ee-53fc84aaae02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9f46a2651e067cd11996fb1c43221a21ab71fdc3ca78a61d9245516f871a5f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dfcc763fe6ab69cab09f09c92ad9a85a4ded49c1443904134f78d1ca0092dae\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:41:30Z\\\",\\\"message\\\":\\\"2025-12-01T08:40:45+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_81306c4a-540c-48bc-b1a5-86f626f67123\\\\n2025-12-01T08:40:45+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_81306c4a-540c-48bc-b1a5-86f626f67123 to /host/opt/cni/bin/\\\\n2025-12-01T08:40:45Z [verbose] multus-daemon started\\\\n2025-12-01T08:40:45Z [verbose] Readiness Indicator file check\\\\n2025-12-01T08:41:30Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zfr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tpwwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:54Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.676752 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90f08b6fa658ce25fbc25d8d44d956cdb6424a0b047205f19d1a6fd6870b1e90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8d27e3678f1f63f55bd1b76668e328f49368d67d4d7338a6326cb73d13e08f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:54Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.689344 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qpr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf3f8dc-c2aa-4e0a-a842-e053dd5bc0b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c42ab88e96b59d9430a69343e06a329447ae0163cb0f5b9ef2bcafa1bf356583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qpr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:54Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.713005 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shkn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8570b876-a13b-4b8f-99c4-4900cfe19a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0afb7d55f395ee39aa245516631b88aab92399973a628f91172c9e410dcc13dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7028c448e28d30b6faeacbcecc5101d67eedde1c3e8a80341cb9688fd47ad859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa0ea361bd2031da8f4d2653d1b81fc839104e8712338c901662fca01d127023\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c37c15efc930fa8e1317fcf1d25a99d2a48c7094db9ddd8166374d83a99ecc1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd8a063a96e33369563219a06faf384b0098048569034987ce164f93abb182e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bad275213db78d56dec1ce0876feb5a06b85979bc2ad73f33874da34ef22fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fcc63acdaf5f71eb24a64bbd7e8054c791d38f5562bc1d8909e14e776f2d8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mbkv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shkn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:54Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.716236 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.716313 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.716332 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.716356 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.716374 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:54Z","lastTransitionTime":"2025-12-01T08:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.728971 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfqcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06182c66-4be6-4e00-9a17-4917badfc997\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa8ef5ef07129ec4a7b4cf4738a01cd6444e9bf548134a9da493c7149301387\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfqcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:54Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.750970 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f7ac0b4-d048-46e4-bcee-4a271503e46d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8208cd5fda4df2e3b5a8bd90a18c8aca62b16cc2c60e06f52411f24f3d60dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba3cfd8a611c687414d7ecd6b02b423b351d7ac2a4d041085d4ca3bb2b0709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ae6257c73f3632c499b5f107178f64a9cac2f4325c57b41aa75d6d4ea14dd12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5088b5bf7bc847f59ca9efbf400855a70b99f1bbbf207eb93576c3b3eabae5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:54Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.767688 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:54Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.781690 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36539974b64191f4ca087bacbd02333f90c89fc8273ab7285e9a1d840155707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:54Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.794992 4873 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef7b114-0e07-402d-a37b-315c36011f4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26e4a6e0a8877a213e2e974d4f7a3ad84938ae1365fa45ca7621bc050470e7e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67d090dbd4e672a6d3e777da244480eaf270518c68f8eaf2292ef7c366b8da4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgsbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-scwpp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:54Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.819707 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.819753 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.819770 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.819788 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:54 crc kubenswrapper[4873]: I1201 08:41:54.819805 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:54Z","lastTransitionTime":"2025-12-01T08:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:55 crc kubenswrapper[4873]: I1201 08:41:55.278310 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:55 crc kubenswrapper[4873]: I1201 08:41:55.278408 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:55 crc kubenswrapper[4873]: I1201 08:41:55.278434 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:55 crc kubenswrapper[4873]: I1201 08:41:55.278933 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:55 crc kubenswrapper[4873]: I1201 08:41:55.278958 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:55Z","lastTransitionTime":"2025-12-01T08:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:55 crc kubenswrapper[4873]: I1201 08:41:55.382696 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:55 crc kubenswrapper[4873]: I1201 08:41:55.382787 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:55 crc kubenswrapper[4873]: I1201 08:41:55.382806 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:55 crc kubenswrapper[4873]: I1201 08:41:55.382832 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:55 crc kubenswrapper[4873]: I1201 08:41:55.382851 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:55Z","lastTransitionTime":"2025-12-01T08:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:55 crc kubenswrapper[4873]: I1201 08:41:55.515174 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:55 crc kubenswrapper[4873]: I1201 08:41:55.515242 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:55 crc kubenswrapper[4873]: I1201 08:41:55.515262 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:55 crc kubenswrapper[4873]: I1201 08:41:55.515287 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:55 crc kubenswrapper[4873]: I1201 08:41:55.515305 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:55Z","lastTransitionTime":"2025-12-01T08:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:55 crc kubenswrapper[4873]: I1201 08:41:55.619508 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:55 crc kubenswrapper[4873]: I1201 08:41:55.619570 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:55 crc kubenswrapper[4873]: I1201 08:41:55.619588 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:55 crc kubenswrapper[4873]: I1201 08:41:55.619612 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:55 crc kubenswrapper[4873]: I1201 08:41:55.619632 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:55Z","lastTransitionTime":"2025-12-01T08:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:55 crc kubenswrapper[4873]: I1201 08:41:55.724301 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:55 crc kubenswrapper[4873]: I1201 08:41:55.724370 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:55 crc kubenswrapper[4873]: I1201 08:41:55.724395 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:55 crc kubenswrapper[4873]: I1201 08:41:55.724425 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:55 crc kubenswrapper[4873]: I1201 08:41:55.724443 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:55Z","lastTransitionTime":"2025-12-01T08:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:55 crc kubenswrapper[4873]: I1201 08:41:55.827347 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:55 crc kubenswrapper[4873]: I1201 08:41:55.827413 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:55 crc kubenswrapper[4873]: I1201 08:41:55.827433 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:55 crc kubenswrapper[4873]: I1201 08:41:55.827458 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:55 crc kubenswrapper[4873]: I1201 08:41:55.827478 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:55Z","lastTransitionTime":"2025-12-01T08:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:55 crc kubenswrapper[4873]: I1201 08:41:55.932392 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:55 crc kubenswrapper[4873]: I1201 08:41:55.932472 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:55 crc kubenswrapper[4873]: I1201 08:41:55.932486 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:55 crc kubenswrapper[4873]: I1201 08:41:55.932508 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:55 crc kubenswrapper[4873]: I1201 08:41:55.932522 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:55Z","lastTransitionTime":"2025-12-01T08:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.036412 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.036474 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.036496 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.036525 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.036547 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:56Z","lastTransitionTime":"2025-12-01T08:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.139543 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.139595 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.139611 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.139631 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.139645 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:56Z","lastTransitionTime":"2025-12-01T08:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.243274 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.243315 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.243328 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.243343 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.243352 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:56Z","lastTransitionTime":"2025-12-01T08:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.346516 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.346558 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.346568 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.346583 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.346592 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:56Z","lastTransitionTime":"2025-12-01T08:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.429957 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.430043 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.429957 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.430146 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:41:56 crc kubenswrapper[4873]: E1201 08:41:56.430288 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:56 crc kubenswrapper[4873]: E1201 08:41:56.430357 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:56 crc kubenswrapper[4873]: E1201 08:41:56.430446 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:56 crc kubenswrapper[4873]: E1201 08:41:56.430716 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.449842 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.449913 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.449937 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.449969 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.449988 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:56Z","lastTransitionTime":"2025-12-01T08:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.554332 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.554389 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.554401 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.554422 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.554435 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:56Z","lastTransitionTime":"2025-12-01T08:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.658514 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.658586 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.658606 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.658632 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.658650 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:56Z","lastTransitionTime":"2025-12-01T08:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.762528 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.762607 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.762631 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.762663 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.762684 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:56Z","lastTransitionTime":"2025-12-01T08:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.866543 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.866621 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.866651 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.866684 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.866708 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:56Z","lastTransitionTime":"2025-12-01T08:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.969382 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.969456 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.969475 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.969503 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:56 crc kubenswrapper[4873]: I1201 08:41:56.969523 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:56Z","lastTransitionTime":"2025-12-01T08:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.073381 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.073471 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.073496 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.073528 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.073548 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:57Z","lastTransitionTime":"2025-12-01T08:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.177398 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.177485 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.177497 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.177511 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.177520 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:57Z","lastTransitionTime":"2025-12-01T08:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.279601 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.279669 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.279694 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.279728 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.279752 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:57Z","lastTransitionTime":"2025-12-01T08:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.379495 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.379564 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.379582 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.379609 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.379627 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:57Z","lastTransitionTime":"2025-12-01T08:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:57 crc kubenswrapper[4873]: E1201 08:41:57.401819 4873 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963d6124-09f6-48b8-b38c-854877aa92e3\\\",\\\"systemUUID\\\":\\\"27c66b0b-2450-400c-b6ea-cdb9cbbf95ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.408006 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.408159 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.408186 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.408218 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.408246 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:57Z","lastTransitionTime":"2025-12-01T08:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:57 crc kubenswrapper[4873]: E1201 08:41:57.429362 4873 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963d6124-09f6-48b8-b38c-854877aa92e3\\\",\\\"systemUUID\\\":\\\"27c66b0b-2450-400c-b6ea-cdb9cbbf95ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.439777 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.439864 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.439892 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.439927 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.439956 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:57Z","lastTransitionTime":"2025-12-01T08:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:57 crc kubenswrapper[4873]: E1201 08:41:57.464299 4873 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963d6124-09f6-48b8-b38c-854877aa92e3\\\",\\\"systemUUID\\\":\\\"27c66b0b-2450-400c-b6ea-cdb9cbbf95ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.470841 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.470906 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.470927 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.470952 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.470970 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:57Z","lastTransitionTime":"2025-12-01T08:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:57 crc kubenswrapper[4873]: E1201 08:41:57.492515 4873 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963d6124-09f6-48b8-b38c-854877aa92e3\\\",\\\"systemUUID\\\":\\\"27c66b0b-2450-400c-b6ea-cdb9cbbf95ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.498501 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.498574 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.498589 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.498613 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.498630 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:57Z","lastTransitionTime":"2025-12-01T08:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:57 crc kubenswrapper[4873]: E1201 08:41:57.518774 4873 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963d6124-09f6-48b8-b38c-854877aa92e3\\\",\\\"systemUUID\\\":\\\"27c66b0b-2450-400c-b6ea-cdb9cbbf95ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:57 crc kubenswrapper[4873]: E1201 08:41:57.518992 4873 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.521852 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.521918 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.521941 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.521972 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.521996 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:57Z","lastTransitionTime":"2025-12-01T08:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.626612 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.626676 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.626694 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.626724 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.626743 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:57Z","lastTransitionTime":"2025-12-01T08:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.730773 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.730864 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.730885 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.730911 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.730934 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:57Z","lastTransitionTime":"2025-12-01T08:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.835292 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.835343 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.835355 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.835373 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.835387 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:57Z","lastTransitionTime":"2025-12-01T08:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.940179 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.940244 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.940261 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.940285 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:57 crc kubenswrapper[4873]: I1201 08:41:57.940302 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:57Z","lastTransitionTime":"2025-12-01T08:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.044277 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.044343 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.044361 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.044387 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.044408 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:58Z","lastTransitionTime":"2025-12-01T08:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.147541 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.147626 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.147651 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.147687 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.147709 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:58Z","lastTransitionTime":"2025-12-01T08:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.250815 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.250871 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.250884 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.250903 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.250915 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:58Z","lastTransitionTime":"2025-12-01T08:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.354258 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.354344 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.354359 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.354384 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.354401 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:58Z","lastTransitionTime":"2025-12-01T08:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.430243 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.430365 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.430243 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:58 crc kubenswrapper[4873]: E1201 08:41:58.430495 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:58 crc kubenswrapper[4873]: E1201 08:41:58.430772 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:58 crc kubenswrapper[4873]: E1201 08:41:58.430896 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.430971 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:41:58 crc kubenswrapper[4873]: E1201 08:41:58.431121 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.457687 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.457778 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.457803 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.457835 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.457861 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:58Z","lastTransitionTime":"2025-12-01T08:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.562209 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.562270 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.562287 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.562312 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.562334 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:58Z","lastTransitionTime":"2025-12-01T08:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.665704 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.665778 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.665797 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.665822 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.665841 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:58Z","lastTransitionTime":"2025-12-01T08:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.769579 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.769674 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.769690 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.769719 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.769739 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:58Z","lastTransitionTime":"2025-12-01T08:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.873373 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.873430 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.873447 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.873469 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.873487 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:58Z","lastTransitionTime":"2025-12-01T08:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.977082 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.977153 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.977171 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.977196 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:58 crc kubenswrapper[4873]: I1201 08:41:58.977221 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:58Z","lastTransitionTime":"2025-12-01T08:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:59 crc kubenswrapper[4873]: I1201 08:41:59.080297 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:59 crc kubenswrapper[4873]: I1201 08:41:59.080781 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:59 crc kubenswrapper[4873]: I1201 08:41:59.080798 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:59 crc kubenswrapper[4873]: I1201 08:41:59.080823 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:59 crc kubenswrapper[4873]: I1201 08:41:59.080840 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:59Z","lastTransitionTime":"2025-12-01T08:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:59 crc kubenswrapper[4873]: I1201 08:41:59.184362 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:59 crc kubenswrapper[4873]: I1201 08:41:59.184433 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:59 crc kubenswrapper[4873]: I1201 08:41:59.184451 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:59 crc kubenswrapper[4873]: I1201 08:41:59.184476 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:59 crc kubenswrapper[4873]: I1201 08:41:59.184494 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:59Z","lastTransitionTime":"2025-12-01T08:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:59 crc kubenswrapper[4873]: I1201 08:41:59.287876 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:59 crc kubenswrapper[4873]: I1201 08:41:59.287940 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:59 crc kubenswrapper[4873]: I1201 08:41:59.287958 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:59 crc kubenswrapper[4873]: I1201 08:41:59.287980 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:59 crc kubenswrapper[4873]: I1201 08:41:59.287998 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:59Z","lastTransitionTime":"2025-12-01T08:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:59 crc kubenswrapper[4873]: I1201 08:41:59.391579 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:59 crc kubenswrapper[4873]: I1201 08:41:59.391685 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:59 crc kubenswrapper[4873]: I1201 08:41:59.391705 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:59 crc kubenswrapper[4873]: I1201 08:41:59.391731 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:59 crc kubenswrapper[4873]: I1201 08:41:59.391751 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:59Z","lastTransitionTime":"2025-12-01T08:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:59 crc kubenswrapper[4873]: I1201 08:41:59.495752 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:59 crc kubenswrapper[4873]: I1201 08:41:59.495846 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:59 crc kubenswrapper[4873]: I1201 08:41:59.495873 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:59 crc kubenswrapper[4873]: I1201 08:41:59.495911 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:59 crc kubenswrapper[4873]: I1201 08:41:59.495939 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:59Z","lastTransitionTime":"2025-12-01T08:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:59 crc kubenswrapper[4873]: I1201 08:41:59.599578 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:59 crc kubenswrapper[4873]: I1201 08:41:59.599642 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:59 crc kubenswrapper[4873]: I1201 08:41:59.599661 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:59 crc kubenswrapper[4873]: I1201 08:41:59.599685 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:59 crc kubenswrapper[4873]: I1201 08:41:59.599708 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:59Z","lastTransitionTime":"2025-12-01T08:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:59 crc kubenswrapper[4873]: I1201 08:41:59.702598 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:59 crc kubenswrapper[4873]: I1201 08:41:59.702666 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:59 crc kubenswrapper[4873]: I1201 08:41:59.702685 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:59 crc kubenswrapper[4873]: I1201 08:41:59.702714 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:59 crc kubenswrapper[4873]: I1201 08:41:59.702735 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:59Z","lastTransitionTime":"2025-12-01T08:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:59 crc kubenswrapper[4873]: I1201 08:41:59.805730 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:59 crc kubenswrapper[4873]: I1201 08:41:59.805803 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:59 crc kubenswrapper[4873]: I1201 08:41:59.805824 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:59 crc kubenswrapper[4873]: I1201 08:41:59.805851 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:59 crc kubenswrapper[4873]: I1201 08:41:59.805870 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:59Z","lastTransitionTime":"2025-12-01T08:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:59 crc kubenswrapper[4873]: I1201 08:41:59.908656 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:59 crc kubenswrapper[4873]: I1201 08:41:59.908739 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:59 crc kubenswrapper[4873]: I1201 08:41:59.908757 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:59 crc kubenswrapper[4873]: I1201 08:41:59.908786 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:59 crc kubenswrapper[4873]: I1201 08:41:59.908831 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:59Z","lastTransitionTime":"2025-12-01T08:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.011419 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.011479 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.011496 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.011522 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.011542 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:00Z","lastTransitionTime":"2025-12-01T08:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.114742 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.114802 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.114821 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.114850 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.114870 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:00Z","lastTransitionTime":"2025-12-01T08:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.218929 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.219059 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.219088 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.219122 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.219147 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:00Z","lastTransitionTime":"2025-12-01T08:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.322416 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.322490 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.322511 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.322539 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.322563 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:00Z","lastTransitionTime":"2025-12-01T08:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.425508 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.425592 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.425619 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.425649 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.425671 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:00Z","lastTransitionTime":"2025-12-01T08:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.429199 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:42:00 crc kubenswrapper[4873]: E1201 08:42:00.429353 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.429639 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:42:00 crc kubenswrapper[4873]: E1201 08:42:00.429754 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.429966 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:42:00 crc kubenswrapper[4873]: E1201 08:42:00.430096 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.430349 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:42:00 crc kubenswrapper[4873]: E1201 08:42:00.430445 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.529071 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.529173 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.529200 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.529230 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.529251 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:00Z","lastTransitionTime":"2025-12-01T08:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.632616 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.632671 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.632688 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.632711 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.632729 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:00Z","lastTransitionTime":"2025-12-01T08:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.739195 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.739242 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.739257 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.739274 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.739286 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:00Z","lastTransitionTime":"2025-12-01T08:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.842892 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.842955 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.842972 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.843003 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.843046 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:00Z","lastTransitionTime":"2025-12-01T08:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.946691 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.946755 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.946773 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.946800 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:00 crc kubenswrapper[4873]: I1201 08:42:00.946822 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:00Z","lastTransitionTime":"2025-12-01T08:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.049497 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.049553 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.049588 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.049619 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.049640 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:01Z","lastTransitionTime":"2025-12-01T08:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.153741 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.153780 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.153790 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.153807 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.153819 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:01Z","lastTransitionTime":"2025-12-01T08:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.258558 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.258634 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.258650 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.258668 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.258682 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:01Z","lastTransitionTime":"2025-12-01T08:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.362077 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.362133 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.362151 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.362174 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.362191 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:01Z","lastTransitionTime":"2025-12-01T08:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.465124 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.465219 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.465234 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.465253 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.465264 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:01Z","lastTransitionTime":"2025-12-01T08:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.567964 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.568036 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.568048 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.568066 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.568077 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:01Z","lastTransitionTime":"2025-12-01T08:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.671759 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.671810 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.671820 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.671838 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.671851 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:01Z","lastTransitionTime":"2025-12-01T08:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.774847 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.774915 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.774927 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.774948 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.774964 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:01Z","lastTransitionTime":"2025-12-01T08:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.878174 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.878234 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.878249 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.878271 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.878285 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:01Z","lastTransitionTime":"2025-12-01T08:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.982079 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.982158 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.982180 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.982207 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:01 crc kubenswrapper[4873]: I1201 08:42:01.982230 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:01Z","lastTransitionTime":"2025-12-01T08:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.084902 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.084990 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.085059 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.085097 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.085121 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:02Z","lastTransitionTime":"2025-12-01T08:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.188888 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.188955 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.188972 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.188997 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.189051 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:02Z","lastTransitionTime":"2025-12-01T08:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.292802 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.292865 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.292876 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.292898 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.292916 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:02Z","lastTransitionTime":"2025-12-01T08:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.396173 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.396300 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.396322 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.396351 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.396369 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:02Z","lastTransitionTime":"2025-12-01T08:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.429885 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.430048 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.430071 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:42:02 crc kubenswrapper[4873]: E1201 08:42:02.430426 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.430530 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:42:02 crc kubenswrapper[4873]: E1201 08:42:02.430606 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:42:02 crc kubenswrapper[4873]: E1201 08:42:02.430718 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:42:02 crc kubenswrapper[4873]: E1201 08:42:02.430850 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.499189 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.499261 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.499281 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.499310 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.499332 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:02Z","lastTransitionTime":"2025-12-01T08:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.602510 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.602590 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.602608 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.602635 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.602658 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:02Z","lastTransitionTime":"2025-12-01T08:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.707115 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.707194 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.707213 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.707251 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.707271 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:02Z","lastTransitionTime":"2025-12-01T08:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.805981 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/08d02e64-a000-4f11-837e-82ad06d925f5-metrics-certs\") pod \"network-metrics-daemon-9vq2r\" (UID: \"08d02e64-a000-4f11-837e-82ad06d925f5\") " pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:42:02 crc kubenswrapper[4873]: E1201 08:42:02.806247 4873 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 08:42:02 crc kubenswrapper[4873]: E1201 08:42:02.806430 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/08d02e64-a000-4f11-837e-82ad06d925f5-metrics-certs podName:08d02e64-a000-4f11-837e-82ad06d925f5 nodeName:}" failed. No retries permitted until 2025-12-01 08:43:06.80639231 +0000 UTC m=+162.708500889 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/08d02e64-a000-4f11-837e-82ad06d925f5-metrics-certs") pod "network-metrics-daemon-9vq2r" (UID: "08d02e64-a000-4f11-837e-82ad06d925f5") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.810219 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.810285 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.810310 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.810377 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.810400 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:02Z","lastTransitionTime":"2025-12-01T08:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.914505 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.914610 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.914650 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.914687 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:02 crc kubenswrapper[4873]: I1201 08:42:02.914706 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:02Z","lastTransitionTime":"2025-12-01T08:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.024415 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.024482 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.024503 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.024558 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.024579 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:03Z","lastTransitionTime":"2025-12-01T08:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.127478 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.127528 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.127540 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.127558 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.127570 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:03Z","lastTransitionTime":"2025-12-01T08:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.231670 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.231753 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.231772 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.231804 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.231825 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:03Z","lastTransitionTime":"2025-12-01T08:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.335229 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.335330 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.335352 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.335383 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.335403 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:03Z","lastTransitionTime":"2025-12-01T08:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.430846 4873 scope.go:117] "RemoveContainer" containerID="c55ab10131e65cc646cead8704fd95f3534c6726b01283d2e3fdf4b8da4e03ff" Dec 01 08:42:03 crc kubenswrapper[4873]: E1201 08:42:03.431141 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dxlvk_openshift-ovn-kubernetes(a308045b-f54d-42ec-8fdf-5bc3e54ef363)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.438871 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.438939 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.438957 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.438984 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.439002 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:03Z","lastTransitionTime":"2025-12-01T08:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.544248 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.544309 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.544324 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.544348 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.544364 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:03Z","lastTransitionTime":"2025-12-01T08:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.647620 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.647716 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.647737 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.647763 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.647783 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:03Z","lastTransitionTime":"2025-12-01T08:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.751000 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.751110 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.751130 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.751158 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.751178 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:03Z","lastTransitionTime":"2025-12-01T08:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.854971 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.855109 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.855136 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.855166 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.855188 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:03Z","lastTransitionTime":"2025-12-01T08:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.958424 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.958492 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.958517 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.958547 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:03 crc kubenswrapper[4873]: I1201 08:42:03.958603 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:03Z","lastTransitionTime":"2025-12-01T08:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.062246 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.062315 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.062333 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.062358 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.062376 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:04Z","lastTransitionTime":"2025-12-01T08:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.165680 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.165760 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.165781 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.165812 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.165836 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:04Z","lastTransitionTime":"2025-12-01T08:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.269357 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.269416 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.269433 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.269456 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.269472 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:04Z","lastTransitionTime":"2025-12-01T08:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.373555 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.373628 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.373649 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.373676 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.373697 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:04Z","lastTransitionTime":"2025-12-01T08:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.429890 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.429991 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:42:04 crc kubenswrapper[4873]: E1201 08:42:04.430284 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.430335 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.430377 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:42:04 crc kubenswrapper[4873]: E1201 08:42:04.430514 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:42:04 crc kubenswrapper[4873]: E1201 08:42:04.430682 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:42:04 crc kubenswrapper[4873]: E1201 08:42:04.430821 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.480179 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.480230 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.480249 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.480317 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.480340 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:04Z","lastTransitionTime":"2025-12-01T08:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.488403 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-qpr6r" podStartSLOduration=81.488380014 podStartE2EDuration="1m21.488380014s" podCreationTimestamp="2025-12-01 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:04.487716936 +0000 UTC m=+100.389825505" watchObservedRunningTime="2025-12-01 08:42:04.488380014 +0000 UTC m=+100.390488583" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.520336 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-shkn9" podStartSLOduration=81.52030592 podStartE2EDuration="1m21.52030592s" podCreationTimestamp="2025-12-01 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:04.519668943 +0000 UTC m=+100.421777512" watchObservedRunningTime="2025-12-01 08:42:04.52030592 +0000 UTC m=+100.422414489" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.542863 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=82.542823794 podStartE2EDuration="1m22.542823794s" podCreationTimestamp="2025-12-01 08:40:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:04.542606879 +0000 UTC m=+100.444715448" watchObservedRunningTime="2025-12-01 08:42:04.542823794 +0000 UTC m=+100.444932373" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.583642 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.583714 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.583733 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.583762 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.583781 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:04Z","lastTransitionTime":"2025-12-01T08:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.607431 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podStartSLOduration=81.607396927 podStartE2EDuration="1m21.607396927s" podCreationTimestamp="2025-12-01 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:04.607196221 +0000 UTC m=+100.509304850" watchObservedRunningTime="2025-12-01 08:42:04.607396927 +0000 UTC m=+100.509505506" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.625974 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-pfqcc" podStartSLOduration=81.625934274 podStartE2EDuration="1m21.625934274s" podCreationTimestamp="2025-12-01 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:04.625315657 +0000 UTC m=+100.527424256" watchObservedRunningTime="2025-12-01 08:42:04.625934274 +0000 UTC m=+100.528042863" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.664448 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=79.664419566 podStartE2EDuration="1m19.664419566s" podCreationTimestamp="2025-12-01 08:40:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:04.661138138 +0000 UTC m=+100.563246767" watchObservedRunningTime="2025-12-01 08:42:04.664419566 +0000 UTC m=+100.566528145" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.686975 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.687092 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.687118 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.687150 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.687172 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:04Z","lastTransitionTime":"2025-12-01T08:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.789969 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.790035 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.790046 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.790059 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.790070 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:04Z","lastTransitionTime":"2025-12-01T08:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.812702 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5smrf" podStartSLOduration=81.812678564 podStartE2EDuration="1m21.812678564s" podCreationTimestamp="2025-12-01 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:04.797749753 +0000 UTC m=+100.699858292" watchObservedRunningTime="2025-12-01 08:42:04.812678564 +0000 UTC m=+100.714787103" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.837273 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=48.837252233 podStartE2EDuration="48.837252233s" podCreationTimestamp="2025-12-01 08:41:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:04.825488317 +0000 UTC m=+100.727596856" watchObservedRunningTime="2025-12-01 08:42:04.837252233 +0000 UTC m=+100.739360772" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.837847 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=25.837841659 podStartE2EDuration="25.837841659s" podCreationTimestamp="2025-12-01 08:41:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:04.836417981 +0000 UTC m=+100.738526520" watchObservedRunningTime="2025-12-01 08:42:04.837841659 +0000 UTC m=+100.739950198" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.851099 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=82.851078894 podStartE2EDuration="1m22.851078894s" podCreationTimestamp="2025-12-01 08:40:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:04.851074444 +0000 UTC m=+100.753183003" watchObservedRunningTime="2025-12-01 08:42:04.851078894 +0000 UTC m=+100.753187433" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.865567 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-tpwwr" podStartSLOduration=81.865547192 podStartE2EDuration="1m21.865547192s" podCreationTimestamp="2025-12-01 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:04.86509602 +0000 UTC m=+100.767204569" watchObservedRunningTime="2025-12-01 08:42:04.865547192 +0000 UTC m=+100.767655731" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.893375 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.893420 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.893432 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.893449 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.893461 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:04Z","lastTransitionTime":"2025-12-01T08:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.997058 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.997127 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.997149 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.997176 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:04 crc kubenswrapper[4873]: I1201 08:42:04.997198 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:04Z","lastTransitionTime":"2025-12-01T08:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:05 crc kubenswrapper[4873]: I1201 08:42:05.101703 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:05 crc kubenswrapper[4873]: I1201 08:42:05.101783 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:05 crc kubenswrapper[4873]: I1201 08:42:05.101803 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:05 crc kubenswrapper[4873]: I1201 08:42:05.101831 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:05 crc kubenswrapper[4873]: I1201 08:42:05.101857 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:05Z","lastTransitionTime":"2025-12-01T08:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:05 crc kubenswrapper[4873]: I1201 08:42:05.204899 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:05 crc kubenswrapper[4873]: I1201 08:42:05.204977 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:05 crc kubenswrapper[4873]: I1201 08:42:05.205003 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:05 crc kubenswrapper[4873]: I1201 08:42:05.205081 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:05 crc kubenswrapper[4873]: I1201 08:42:05.205106 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:05Z","lastTransitionTime":"2025-12-01T08:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:05 crc kubenswrapper[4873]: I1201 08:42:05.308719 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:05 crc kubenswrapper[4873]: I1201 08:42:05.308813 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:05 crc kubenswrapper[4873]: I1201 08:42:05.308840 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:05 crc kubenswrapper[4873]: I1201 08:42:05.308875 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:05 crc kubenswrapper[4873]: I1201 08:42:05.308900 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:05Z","lastTransitionTime":"2025-12-01T08:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:05 crc kubenswrapper[4873]: I1201 08:42:05.412088 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:05 crc kubenswrapper[4873]: I1201 08:42:05.412167 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:05 crc kubenswrapper[4873]: I1201 08:42:05.412188 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:05 crc kubenswrapper[4873]: I1201 08:42:05.412214 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:05 crc kubenswrapper[4873]: I1201 08:42:05.412234 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:05Z","lastTransitionTime":"2025-12-01T08:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:05 crc kubenswrapper[4873]: I1201 08:42:05.515448 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:05 crc kubenswrapper[4873]: I1201 08:42:05.515511 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:05 crc kubenswrapper[4873]: I1201 08:42:05.515528 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:05 crc kubenswrapper[4873]: I1201 08:42:05.515551 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:05 crc kubenswrapper[4873]: I1201 08:42:05.515565 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:05Z","lastTransitionTime":"2025-12-01T08:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:05 crc kubenswrapper[4873]: I1201 08:42:05.619476 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:05 crc kubenswrapper[4873]: I1201 08:42:05.619548 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:05 crc kubenswrapper[4873]: I1201 08:42:05.619563 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:05 crc kubenswrapper[4873]: I1201 08:42:05.619589 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:05 crc kubenswrapper[4873]: I1201 08:42:05.619606 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:05Z","lastTransitionTime":"2025-12-01T08:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:05 crc kubenswrapper[4873]: I1201 08:42:05.724101 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:05 crc kubenswrapper[4873]: I1201 08:42:05.724186 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:05 crc kubenswrapper[4873]: I1201 08:42:05.724210 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:05 crc kubenswrapper[4873]: I1201 08:42:05.724244 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:05 crc kubenswrapper[4873]: I1201 08:42:05.724267 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:05Z","lastTransitionTime":"2025-12-01T08:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:05 crc kubenswrapper[4873]: I1201 08:42:05.828712 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:05 crc kubenswrapper[4873]: I1201 08:42:05.828794 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:05 crc kubenswrapper[4873]: I1201 08:42:05.828813 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:05 crc kubenswrapper[4873]: I1201 08:42:05.828844 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:05 crc kubenswrapper[4873]: I1201 08:42:05.828865 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:05Z","lastTransitionTime":"2025-12-01T08:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:05 crc kubenswrapper[4873]: I1201 08:42:05.932717 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:05 crc kubenswrapper[4873]: I1201 08:42:05.932802 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:05 crc kubenswrapper[4873]: I1201 08:42:05.932823 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:05 crc kubenswrapper[4873]: I1201 08:42:05.932853 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:05 crc kubenswrapper[4873]: I1201 08:42:05.932873 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:05Z","lastTransitionTime":"2025-12-01T08:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.035639 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.035701 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.035720 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.035745 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.035764 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:06Z","lastTransitionTime":"2025-12-01T08:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.138997 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.139074 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.139085 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.139105 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.139114 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:06Z","lastTransitionTime":"2025-12-01T08:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.242619 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.242677 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.242685 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.242706 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.242716 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:06Z","lastTransitionTime":"2025-12-01T08:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.345957 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.346053 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.346075 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.346101 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.346121 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:06Z","lastTransitionTime":"2025-12-01T08:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.429966 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.430066 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.430137 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:42:06 crc kubenswrapper[4873]: E1201 08:42:06.430267 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.430367 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:42:06 crc kubenswrapper[4873]: E1201 08:42:06.430524 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:42:06 crc kubenswrapper[4873]: E1201 08:42:06.430694 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:42:06 crc kubenswrapper[4873]: E1201 08:42:06.430835 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.449556 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.449635 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.449657 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.449684 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.449705 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:06Z","lastTransitionTime":"2025-12-01T08:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.552439 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.552481 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.552496 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.552516 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.552529 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:06Z","lastTransitionTime":"2025-12-01T08:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.656599 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.656687 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.656713 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.656747 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.656771 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:06Z","lastTransitionTime":"2025-12-01T08:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.760618 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.760686 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.760703 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.760781 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.760812 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:06Z","lastTransitionTime":"2025-12-01T08:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.864550 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.864630 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.864652 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.864678 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.864700 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:06Z","lastTransitionTime":"2025-12-01T08:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.969819 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.969904 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.969941 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.969971 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:06 crc kubenswrapper[4873]: I1201 08:42:06.969993 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:06Z","lastTransitionTime":"2025-12-01T08:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.073600 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.073713 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.073732 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.073756 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.073775 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:07Z","lastTransitionTime":"2025-12-01T08:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.177430 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.177505 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.177521 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.177538 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.177550 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:07Z","lastTransitionTime":"2025-12-01T08:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.285620 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.285734 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.285766 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.285801 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.285836 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:07Z","lastTransitionTime":"2025-12-01T08:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.389633 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.389698 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.389707 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.389723 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.389734 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:07Z","lastTransitionTime":"2025-12-01T08:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.502180 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.502229 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.502242 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.502267 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.502285 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:07Z","lastTransitionTime":"2025-12-01T08:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.606880 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.606962 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.606984 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.607044 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.607063 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:07Z","lastTransitionTime":"2025-12-01T08:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.710377 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.710451 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.710468 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.710500 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.710520 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:07Z","lastTransitionTime":"2025-12-01T08:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.745043 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.745088 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.745105 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.745131 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.745151 4873 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:42:07Z","lastTransitionTime":"2025-12-01T08:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.806100 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-fnscv"] Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.806695 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fnscv" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.808856 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.809648 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.809803 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.813736 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.866432 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/3e05a4fd-db9e-44e3-9fd0-2097aefdfee5-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-fnscv\" (UID: \"3e05a4fd-db9e-44e3-9fd0-2097aefdfee5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fnscv" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.866527 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3e05a4fd-db9e-44e3-9fd0-2097aefdfee5-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-fnscv\" (UID: \"3e05a4fd-db9e-44e3-9fd0-2097aefdfee5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fnscv" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.866575 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/3e05a4fd-db9e-44e3-9fd0-2097aefdfee5-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-fnscv\" (UID: \"3e05a4fd-db9e-44e3-9fd0-2097aefdfee5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fnscv" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.866724 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3e05a4fd-db9e-44e3-9fd0-2097aefdfee5-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-fnscv\" (UID: \"3e05a4fd-db9e-44e3-9fd0-2097aefdfee5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fnscv" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.866870 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3e05a4fd-db9e-44e3-9fd0-2097aefdfee5-service-ca\") pod \"cluster-version-operator-5c965bbfc6-fnscv\" (UID: \"3e05a4fd-db9e-44e3-9fd0-2097aefdfee5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fnscv" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.968218 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3e05a4fd-db9e-44e3-9fd0-2097aefdfee5-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-fnscv\" (UID: \"3e05a4fd-db9e-44e3-9fd0-2097aefdfee5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fnscv" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.968287 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/3e05a4fd-db9e-44e3-9fd0-2097aefdfee5-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-fnscv\" (UID: \"3e05a4fd-db9e-44e3-9fd0-2097aefdfee5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fnscv" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.968350 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3e05a4fd-db9e-44e3-9fd0-2097aefdfee5-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-fnscv\" (UID: \"3e05a4fd-db9e-44e3-9fd0-2097aefdfee5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fnscv" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.968432 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3e05a4fd-db9e-44e3-9fd0-2097aefdfee5-service-ca\") pod \"cluster-version-operator-5c965bbfc6-fnscv\" (UID: \"3e05a4fd-db9e-44e3-9fd0-2097aefdfee5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fnscv" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.968467 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/3e05a4fd-db9e-44e3-9fd0-2097aefdfee5-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-fnscv\" (UID: \"3e05a4fd-db9e-44e3-9fd0-2097aefdfee5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fnscv" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.968528 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/3e05a4fd-db9e-44e3-9fd0-2097aefdfee5-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-fnscv\" (UID: \"3e05a4fd-db9e-44e3-9fd0-2097aefdfee5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fnscv" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.968666 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/3e05a4fd-db9e-44e3-9fd0-2097aefdfee5-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-fnscv\" (UID: \"3e05a4fd-db9e-44e3-9fd0-2097aefdfee5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fnscv" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.970493 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3e05a4fd-db9e-44e3-9fd0-2097aefdfee5-service-ca\") pod \"cluster-version-operator-5c965bbfc6-fnscv\" (UID: \"3e05a4fd-db9e-44e3-9fd0-2097aefdfee5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fnscv" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.981528 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3e05a4fd-db9e-44e3-9fd0-2097aefdfee5-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-fnscv\" (UID: \"3e05a4fd-db9e-44e3-9fd0-2097aefdfee5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fnscv" Dec 01 08:42:07 crc kubenswrapper[4873]: I1201 08:42:07.997480 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3e05a4fd-db9e-44e3-9fd0-2097aefdfee5-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-fnscv\" (UID: \"3e05a4fd-db9e-44e3-9fd0-2097aefdfee5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fnscv" Dec 01 08:42:08 crc kubenswrapper[4873]: I1201 08:42:08.133618 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fnscv" Dec 01 08:42:08 crc kubenswrapper[4873]: W1201 08:42:08.158854 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3e05a4fd_db9e_44e3_9fd0_2097aefdfee5.slice/crio-f665a6784dfa6267184ec6e478bd440706ce92556009b10c310294caa7b20b08 WatchSource:0}: Error finding container f665a6784dfa6267184ec6e478bd440706ce92556009b10c310294caa7b20b08: Status 404 returned error can't find the container with id f665a6784dfa6267184ec6e478bd440706ce92556009b10c310294caa7b20b08 Dec 01 08:42:08 crc kubenswrapper[4873]: I1201 08:42:08.330550 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fnscv" event={"ID":"3e05a4fd-db9e-44e3-9fd0-2097aefdfee5","Type":"ContainerStarted","Data":"ab2c5dd3972578553555eb0ab55daffb010d6ff947782182b88f1f965bd6da9f"} Dec 01 08:42:08 crc kubenswrapper[4873]: I1201 08:42:08.330647 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fnscv" event={"ID":"3e05a4fd-db9e-44e3-9fd0-2097aefdfee5","Type":"ContainerStarted","Data":"f665a6784dfa6267184ec6e478bd440706ce92556009b10c310294caa7b20b08"} Dec 01 08:42:08 crc kubenswrapper[4873]: I1201 08:42:08.356867 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fnscv" podStartSLOduration=85.356831804 podStartE2EDuration="1m25.356831804s" podCreationTimestamp="2025-12-01 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:08.354376818 +0000 UTC m=+104.256485387" watchObservedRunningTime="2025-12-01 08:42:08.356831804 +0000 UTC m=+104.258940383" Dec 01 08:42:08 crc kubenswrapper[4873]: I1201 08:42:08.429913 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:42:08 crc kubenswrapper[4873]: I1201 08:42:08.429946 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:42:08 crc kubenswrapper[4873]: I1201 08:42:08.430167 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:42:08 crc kubenswrapper[4873]: I1201 08:42:08.430417 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:42:08 crc kubenswrapper[4873]: E1201 08:42:08.430613 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:42:08 crc kubenswrapper[4873]: E1201 08:42:08.430768 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:42:08 crc kubenswrapper[4873]: E1201 08:42:08.431210 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:42:08 crc kubenswrapper[4873]: E1201 08:42:08.431274 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:42:10 crc kubenswrapper[4873]: I1201 08:42:10.429775 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:42:10 crc kubenswrapper[4873]: I1201 08:42:10.429858 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:42:10 crc kubenswrapper[4873]: I1201 08:42:10.429921 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:42:10 crc kubenswrapper[4873]: I1201 08:42:10.429938 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:42:10 crc kubenswrapper[4873]: E1201 08:42:10.430100 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:42:10 crc kubenswrapper[4873]: E1201 08:42:10.430212 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:42:10 crc kubenswrapper[4873]: E1201 08:42:10.430419 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:42:10 crc kubenswrapper[4873]: E1201 08:42:10.430539 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:42:12 crc kubenswrapper[4873]: I1201 08:42:12.429380 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:42:12 crc kubenswrapper[4873]: I1201 08:42:12.429508 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:42:12 crc kubenswrapper[4873]: I1201 08:42:12.429404 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:42:12 crc kubenswrapper[4873]: I1201 08:42:12.429396 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:42:12 crc kubenswrapper[4873]: E1201 08:42:12.429634 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:42:12 crc kubenswrapper[4873]: E1201 08:42:12.429733 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:42:12 crc kubenswrapper[4873]: E1201 08:42:12.430084 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:42:12 crc kubenswrapper[4873]: E1201 08:42:12.430117 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:42:14 crc kubenswrapper[4873]: I1201 08:42:14.429461 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:42:14 crc kubenswrapper[4873]: I1201 08:42:14.429524 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:42:14 crc kubenswrapper[4873]: I1201 08:42:14.429578 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:42:14 crc kubenswrapper[4873]: I1201 08:42:14.429524 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:42:14 crc kubenswrapper[4873]: E1201 08:42:14.431086 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:42:14 crc kubenswrapper[4873]: E1201 08:42:14.431214 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:42:14 crc kubenswrapper[4873]: E1201 08:42:14.431320 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:42:14 crc kubenswrapper[4873]: E1201 08:42:14.431507 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:42:16 crc kubenswrapper[4873]: I1201 08:42:16.429952 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:42:16 crc kubenswrapper[4873]: E1201 08:42:16.431350 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:42:16 crc kubenswrapper[4873]: I1201 08:42:16.430153 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:42:16 crc kubenswrapper[4873]: E1201 08:42:16.431513 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:42:16 crc kubenswrapper[4873]: I1201 08:42:16.430186 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:42:16 crc kubenswrapper[4873]: E1201 08:42:16.431658 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:42:16 crc kubenswrapper[4873]: I1201 08:42:16.430080 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:42:16 crc kubenswrapper[4873]: E1201 08:42:16.431792 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:42:17 crc kubenswrapper[4873]: I1201 08:42:17.365513 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-tpwwr_13a2246b-93bb-4586-98ee-53fc84aaae02/kube-multus/1.log" Dec 01 08:42:17 crc kubenswrapper[4873]: I1201 08:42:17.366508 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-tpwwr_13a2246b-93bb-4586-98ee-53fc84aaae02/kube-multus/0.log" Dec 01 08:42:17 crc kubenswrapper[4873]: I1201 08:42:17.366596 4873 generic.go:334] "Generic (PLEG): container finished" podID="13a2246b-93bb-4586-98ee-53fc84aaae02" containerID="a9f46a2651e067cd11996fb1c43221a21ab71fdc3ca78a61d9245516f871a5f3" exitCode=1 Dec 01 08:42:17 crc kubenswrapper[4873]: I1201 08:42:17.366647 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-tpwwr" event={"ID":"13a2246b-93bb-4586-98ee-53fc84aaae02","Type":"ContainerDied","Data":"a9f46a2651e067cd11996fb1c43221a21ab71fdc3ca78a61d9245516f871a5f3"} Dec 01 08:42:17 crc kubenswrapper[4873]: I1201 08:42:17.366705 4873 scope.go:117] "RemoveContainer" containerID="3dfcc763fe6ab69cab09f09c92ad9a85a4ded49c1443904134f78d1ca0092dae" Dec 01 08:42:17 crc kubenswrapper[4873]: I1201 08:42:17.367353 4873 scope.go:117] "RemoveContainer" containerID="a9f46a2651e067cd11996fb1c43221a21ab71fdc3ca78a61d9245516f871a5f3" Dec 01 08:42:17 crc kubenswrapper[4873]: E1201 08:42:17.367673 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-tpwwr_openshift-multus(13a2246b-93bb-4586-98ee-53fc84aaae02)\"" pod="openshift-multus/multus-tpwwr" podUID="13a2246b-93bb-4586-98ee-53fc84aaae02" Dec 01 08:42:18 crc kubenswrapper[4873]: I1201 08:42:18.373543 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-tpwwr_13a2246b-93bb-4586-98ee-53fc84aaae02/kube-multus/1.log" Dec 01 08:42:18 crc kubenswrapper[4873]: I1201 08:42:18.429973 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:42:18 crc kubenswrapper[4873]: I1201 08:42:18.430086 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:42:18 crc kubenswrapper[4873]: I1201 08:42:18.429923 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:42:18 crc kubenswrapper[4873]: E1201 08:42:18.430825 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:42:18 crc kubenswrapper[4873]: I1201 08:42:18.430895 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:42:18 crc kubenswrapper[4873]: E1201 08:42:18.431197 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:42:18 crc kubenswrapper[4873]: E1201 08:42:18.431312 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:42:18 crc kubenswrapper[4873]: E1201 08:42:18.431494 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:42:18 crc kubenswrapper[4873]: I1201 08:42:18.432723 4873 scope.go:117] "RemoveContainer" containerID="c55ab10131e65cc646cead8704fd95f3534c6726b01283d2e3fdf4b8da4e03ff" Dec 01 08:42:19 crc kubenswrapper[4873]: I1201 08:42:19.381706 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dxlvk_a308045b-f54d-42ec-8fdf-5bc3e54ef363/ovnkube-controller/3.log" Dec 01 08:42:19 crc kubenswrapper[4873]: I1201 08:42:19.384314 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-9vq2r"] Dec 01 08:42:19 crc kubenswrapper[4873]: I1201 08:42:19.386924 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" event={"ID":"a308045b-f54d-42ec-8fdf-5bc3e54ef363","Type":"ContainerStarted","Data":"d4509219c44fea10d7c78878d40843169c63e790756210df070ae2c8da444cfe"} Dec 01 08:42:19 crc kubenswrapper[4873]: I1201 08:42:19.386948 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:42:19 crc kubenswrapper[4873]: E1201 08:42:19.387314 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:42:19 crc kubenswrapper[4873]: I1201 08:42:19.387695 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:42:19 crc kubenswrapper[4873]: I1201 08:42:19.439777 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" podStartSLOduration=96.439756317 podStartE2EDuration="1m36.439756317s" podCreationTimestamp="2025-12-01 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:19.436977553 +0000 UTC m=+115.339086102" watchObservedRunningTime="2025-12-01 08:42:19.439756317 +0000 UTC m=+115.341864866" Dec 01 08:42:20 crc kubenswrapper[4873]: I1201 08:42:20.429563 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:42:20 crc kubenswrapper[4873]: I1201 08:42:20.429619 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:42:20 crc kubenswrapper[4873]: I1201 08:42:20.429679 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:42:20 crc kubenswrapper[4873]: E1201 08:42:20.429843 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:42:20 crc kubenswrapper[4873]: E1201 08:42:20.430003 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:42:20 crc kubenswrapper[4873]: E1201 08:42:20.430093 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:42:21 crc kubenswrapper[4873]: I1201 08:42:21.429695 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:42:21 crc kubenswrapper[4873]: E1201 08:42:21.430317 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:42:22 crc kubenswrapper[4873]: I1201 08:42:22.429852 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:42:22 crc kubenswrapper[4873]: I1201 08:42:22.430009 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:42:22 crc kubenswrapper[4873]: I1201 08:42:22.430095 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:42:22 crc kubenswrapper[4873]: E1201 08:42:22.430275 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:42:22 crc kubenswrapper[4873]: E1201 08:42:22.430505 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:42:22 crc kubenswrapper[4873]: E1201 08:42:22.430894 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:42:23 crc kubenswrapper[4873]: I1201 08:42:23.429399 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:42:23 crc kubenswrapper[4873]: E1201 08:42:23.429648 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:42:24 crc kubenswrapper[4873]: E1201 08:42:24.408781 4873 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 01 08:42:24 crc kubenswrapper[4873]: I1201 08:42:24.429619 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:42:24 crc kubenswrapper[4873]: I1201 08:42:24.429648 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:42:24 crc kubenswrapper[4873]: E1201 08:42:24.432083 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:42:24 crc kubenswrapper[4873]: I1201 08:42:24.432201 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:42:24 crc kubenswrapper[4873]: E1201 08:42:24.433342 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:42:24 crc kubenswrapper[4873]: E1201 08:42:24.436235 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:42:24 crc kubenswrapper[4873]: E1201 08:42:24.576429 4873 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 01 08:42:25 crc kubenswrapper[4873]: I1201 08:42:25.429153 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:42:25 crc kubenswrapper[4873]: E1201 08:42:25.429353 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:42:26 crc kubenswrapper[4873]: I1201 08:42:26.429155 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:42:26 crc kubenswrapper[4873]: I1201 08:42:26.429214 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:42:26 crc kubenswrapper[4873]: I1201 08:42:26.429285 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:42:26 crc kubenswrapper[4873]: E1201 08:42:26.429336 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:42:26 crc kubenswrapper[4873]: E1201 08:42:26.429526 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:42:26 crc kubenswrapper[4873]: E1201 08:42:26.429665 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:42:27 crc kubenswrapper[4873]: I1201 08:42:27.429289 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:42:27 crc kubenswrapper[4873]: E1201 08:42:27.429546 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:42:28 crc kubenswrapper[4873]: I1201 08:42:28.429561 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:42:28 crc kubenswrapper[4873]: I1201 08:42:28.429670 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:42:28 crc kubenswrapper[4873]: E1201 08:42:28.429782 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:42:28 crc kubenswrapper[4873]: I1201 08:42:28.429867 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:42:28 crc kubenswrapper[4873]: E1201 08:42:28.429977 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:42:28 crc kubenswrapper[4873]: E1201 08:42:28.430183 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:42:29 crc kubenswrapper[4873]: I1201 08:42:29.429357 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:42:29 crc kubenswrapper[4873]: E1201 08:42:29.429549 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:42:29 crc kubenswrapper[4873]: E1201 08:42:29.578382 4873 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 01 08:42:30 crc kubenswrapper[4873]: I1201 08:42:30.429166 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:42:30 crc kubenswrapper[4873]: I1201 08:42:30.429864 4873 scope.go:117] "RemoveContainer" containerID="a9f46a2651e067cd11996fb1c43221a21ab71fdc3ca78a61d9245516f871a5f3" Dec 01 08:42:30 crc kubenswrapper[4873]: I1201 08:42:30.429239 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:42:30 crc kubenswrapper[4873]: E1201 08:42:30.430000 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:42:30 crc kubenswrapper[4873]: I1201 08:42:30.429405 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:42:30 crc kubenswrapper[4873]: E1201 08:42:30.430162 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:42:30 crc kubenswrapper[4873]: E1201 08:42:30.430299 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:42:31 crc kubenswrapper[4873]: I1201 08:42:31.429870 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:42:31 crc kubenswrapper[4873]: E1201 08:42:31.430169 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:42:31 crc kubenswrapper[4873]: I1201 08:42:31.437880 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-tpwwr_13a2246b-93bb-4586-98ee-53fc84aaae02/kube-multus/1.log" Dec 01 08:42:31 crc kubenswrapper[4873]: I1201 08:42:31.437943 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-tpwwr" event={"ID":"13a2246b-93bb-4586-98ee-53fc84aaae02","Type":"ContainerStarted","Data":"b80d1c593405c5217176f034e50359d6fea16f79b5f8758b7753809b6ab96f76"} Dec 01 08:42:32 crc kubenswrapper[4873]: I1201 08:42:32.429111 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:42:32 crc kubenswrapper[4873]: I1201 08:42:32.429204 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:42:32 crc kubenswrapper[4873]: I1201 08:42:32.429111 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:42:32 crc kubenswrapper[4873]: E1201 08:42:32.429373 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:42:32 crc kubenswrapper[4873]: E1201 08:42:32.429540 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:42:32 crc kubenswrapper[4873]: E1201 08:42:32.429740 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:42:33 crc kubenswrapper[4873]: I1201 08:42:33.429879 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:42:33 crc kubenswrapper[4873]: E1201 08:42:33.430125 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9vq2r" podUID="08d02e64-a000-4f11-837e-82ad06d925f5" Dec 01 08:42:34 crc kubenswrapper[4873]: I1201 08:42:34.429905 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:42:34 crc kubenswrapper[4873]: E1201 08:42:34.431569 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:42:34 crc kubenswrapper[4873]: I1201 08:42:34.431673 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:42:34 crc kubenswrapper[4873]: E1201 08:42:34.431920 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:42:34 crc kubenswrapper[4873]: I1201 08:42:34.432229 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:42:34 crc kubenswrapper[4873]: E1201 08:42:34.432341 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:42:35 crc kubenswrapper[4873]: I1201 08:42:35.429890 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:42:35 crc kubenswrapper[4873]: I1201 08:42:35.434358 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 01 08:42:35 crc kubenswrapper[4873]: I1201 08:42:35.434532 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 01 08:42:36 crc kubenswrapper[4873]: I1201 08:42:36.429930 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:42:36 crc kubenswrapper[4873]: I1201 08:42:36.430205 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:42:36 crc kubenswrapper[4873]: I1201 08:42:36.430330 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:42:36 crc kubenswrapper[4873]: I1201 08:42:36.432881 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 01 08:42:36 crc kubenswrapper[4873]: I1201 08:42:36.433058 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 01 08:42:36 crc kubenswrapper[4873]: I1201 08:42:36.434601 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 01 08:42:36 crc kubenswrapper[4873]: I1201 08:42:36.434928 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.147035 4873 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.198108 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-lcqht"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.199205 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lcqht" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.199940 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-52pjn"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.200656 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-52pjn" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.202285 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.202331 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.202457 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-kw4m8"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.202949 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-kw4m8" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.211684 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.211952 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.212174 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.212263 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.212516 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.212615 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.212804 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.212898 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.213118 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.213406 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.213667 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.214165 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.214222 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.214924 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.215090 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.215097 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.215113 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.217925 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.218116 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.218114 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-q4wmp"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.218684 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-w8xn2"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.219420 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tdwnv"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.219851 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gscbx"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.220292 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cjn5c"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.220852 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.221499 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q4wmp" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.221945 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-w8xn2" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.222039 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7dcf8"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.222384 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-xt48p"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.222610 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tdwnv" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.222707 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-xt48p" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.222776 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7dcf8" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.222811 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gscbx" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.222974 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-5bhp6"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.223671 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-5bhp6" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.248584 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.248794 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.249302 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.267249 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.268212 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.269599 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-z6qdr"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.270651 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-nd5kp"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.271158 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-wq2zz"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.271329 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.271663 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-sg6bg"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.272162 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.272281 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-76drh"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.272785 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nf47\" (UniqueName: \"kubernetes.io/projected/40a9c91a-1a6d-438d-9c65-51dc38b34f34-kube-api-access-8nf47\") pod \"controller-manager-879f6c89f-kw4m8\" (UID: \"40a9c91a-1a6d-438d-9c65-51dc38b34f34\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kw4m8" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.272817 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/01924913-40f0-4502-b151-36079ee0ea8c-audit-dir\") pod \"apiserver-7bbb656c7d-lcqht\" (UID: \"01924913-40f0-4502-b151-36079ee0ea8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lcqht" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.272852 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40a9c91a-1a6d-438d-9c65-51dc38b34f34-serving-cert\") pod \"controller-manager-879f6c89f-kw4m8\" (UID: \"40a9c91a-1a6d-438d-9c65-51dc38b34f34\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kw4m8" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.272908 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.272983 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-nd5kp" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.272910 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/fb8fcfca-0933-4a74-85de-859fda0153d6-images\") pod \"machine-api-operator-5694c8668f-52pjn\" (UID: \"fb8fcfca-0933-4a74-85de-859fda0153d6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-52pjn" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.273064 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-z6qdr" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.273080 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x99bn\" (UniqueName: \"kubernetes.io/projected/fb8fcfca-0933-4a74-85de-859fda0153d6-kube-api-access-x99bn\") pod \"machine-api-operator-5694c8668f-52pjn\" (UID: \"fb8fcfca-0933-4a74-85de-859fda0153d6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-52pjn" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.273097 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wq2zz" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.273109 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/fb8fcfca-0933-4a74-85de-859fda0153d6-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-52pjn\" (UID: \"fb8fcfca-0933-4a74-85de-859fda0153d6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-52pjn" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.273132 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/01924913-40f0-4502-b151-36079ee0ea8c-encryption-config\") pod \"apiserver-7bbb656c7d-lcqht\" (UID: \"01924913-40f0-4502-b151-36079ee0ea8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lcqht" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.273164 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/01924913-40f0-4502-b151-36079ee0ea8c-etcd-client\") pod \"apiserver-7bbb656c7d-lcqht\" (UID: \"01924913-40f0-4502-b151-36079ee0ea8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lcqht" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.273183 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/40a9c91a-1a6d-438d-9c65-51dc38b34f34-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-kw4m8\" (UID: \"40a9c91a-1a6d-438d-9c65-51dc38b34f34\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kw4m8" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.273214 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb8fcfca-0933-4a74-85de-859fda0153d6-config\") pod \"machine-api-operator-5694c8668f-52pjn\" (UID: \"fb8fcfca-0933-4a74-85de-859fda0153d6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-52pjn" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.273234 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/01924913-40f0-4502-b151-36079ee0ea8c-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-lcqht\" (UID: \"01924913-40f0-4502-b151-36079ee0ea8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lcqht" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.273257 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40a9c91a-1a6d-438d-9c65-51dc38b34f34-config\") pod \"controller-manager-879f6c89f-kw4m8\" (UID: \"40a9c91a-1a6d-438d-9c65-51dc38b34f34\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kw4m8" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.273277 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdzlb\" (UniqueName: \"kubernetes.io/projected/01924913-40f0-4502-b151-36079ee0ea8c-kube-api-access-tdzlb\") pod \"apiserver-7bbb656c7d-lcqht\" (UID: \"01924913-40f0-4502-b151-36079ee0ea8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lcqht" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.273308 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/01924913-40f0-4502-b151-36079ee0ea8c-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-lcqht\" (UID: \"01924913-40f0-4502-b151-36079ee0ea8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lcqht" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.273330 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/01924913-40f0-4502-b151-36079ee0ea8c-audit-policies\") pod \"apiserver-7bbb656c7d-lcqht\" (UID: \"01924913-40f0-4502-b151-36079ee0ea8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lcqht" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.273351 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/40a9c91a-1a6d-438d-9c65-51dc38b34f34-client-ca\") pod \"controller-manager-879f6c89f-kw4m8\" (UID: \"40a9c91a-1a6d-438d-9c65-51dc38b34f34\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kw4m8" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.273370 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01924913-40f0-4502-b151-36079ee0ea8c-serving-cert\") pod \"apiserver-7bbb656c7d-lcqht\" (UID: \"01924913-40f0-4502-b151-36079ee0ea8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lcqht" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.273586 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.273686 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.274655 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-76drh" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.276240 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.276372 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.276438 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.276473 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.276678 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.276735 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.276866 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.276946 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.276881 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.277127 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.277208 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.277220 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.277227 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.277256 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.277346 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.277376 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.277385 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.277463 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.277497 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.277806 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.279084 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xfb2s"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.279839 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xfb2s" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.280104 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.280291 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.280377 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xg89d"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.280390 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.280548 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.280970 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xg89d" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.281385 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.281549 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ptnr8"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.281886 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.282167 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ptnr8" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.282171 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.282423 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.282272 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.282370 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.282643 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.282889 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.282940 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.283568 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.283733 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.283922 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.284403 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.285517 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.286665 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.286741 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.286874 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.286975 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.287064 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.287088 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.287243 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-cl5cd"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.288959 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-292nl"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.290178 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-292nl" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.288970 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cl5cd" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.291328 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.291496 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.294244 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.294975 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.294991 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.295219 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.295444 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.295720 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.296418 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-qvzhc"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.297459 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-qvzhc" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.297989 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-gxhcq"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.299783 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gxhcq" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.331964 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.333343 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7hstm"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.333528 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.345031 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xdpgx"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.345395 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7hstm" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.345898 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xdpgx" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.346164 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-h9kpk"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.347032 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-h9kpk" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.355797 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.374812 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb8fcfca-0933-4a74-85de-859fda0153d6-config\") pod \"machine-api-operator-5694c8668f-52pjn\" (UID: \"fb8fcfca-0933-4a74-85de-859fda0153d6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-52pjn" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.374869 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/01924913-40f0-4502-b151-36079ee0ea8c-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-lcqht\" (UID: \"01924913-40f0-4502-b151-36079ee0ea8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lcqht" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.374893 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40a9c91a-1a6d-438d-9c65-51dc38b34f34-config\") pod \"controller-manager-879f6c89f-kw4m8\" (UID: \"40a9c91a-1a6d-438d-9c65-51dc38b34f34\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kw4m8" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.374912 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdzlb\" (UniqueName: \"kubernetes.io/projected/01924913-40f0-4502-b151-36079ee0ea8c-kube-api-access-tdzlb\") pod \"apiserver-7bbb656c7d-lcqht\" (UID: \"01924913-40f0-4502-b151-36079ee0ea8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lcqht" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.374940 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjzcq\" (UniqueName: \"kubernetes.io/projected/dc0adee4-65bf-484b-8f55-3c83022cae82-kube-api-access-cjzcq\") pod \"downloads-7954f5f757-76drh\" (UID: \"dc0adee4-65bf-484b-8f55-3c83022cae82\") " pod="openshift-console/downloads-7954f5f757-76drh" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.374973 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/01924913-40f0-4502-b151-36079ee0ea8c-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-lcqht\" (UID: \"01924913-40f0-4502-b151-36079ee0ea8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lcqht" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.374994 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swdsl\" (UniqueName: \"kubernetes.io/projected/1ea01a0b-ae00-469d-a56d-0103ac067d1b-kube-api-access-swdsl\") pod \"authentication-operator-69f744f599-z6qdr\" (UID: \"1ea01a0b-ae00-469d-a56d-0103ac067d1b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-z6qdr" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.375036 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/01924913-40f0-4502-b151-36079ee0ea8c-audit-policies\") pod \"apiserver-7bbb656c7d-lcqht\" (UID: \"01924913-40f0-4502-b151-36079ee0ea8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lcqht" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.375062 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/40a9c91a-1a6d-438d-9c65-51dc38b34f34-client-ca\") pod \"controller-manager-879f6c89f-kw4m8\" (UID: \"40a9c91a-1a6d-438d-9c65-51dc38b34f34\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kw4m8" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.375093 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01924913-40f0-4502-b151-36079ee0ea8c-serving-cert\") pod \"apiserver-7bbb656c7d-lcqht\" (UID: \"01924913-40f0-4502-b151-36079ee0ea8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lcqht" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.375117 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/54e431c7-f38b-4538-a9f6-69d7a8995dce-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tdwnv\" (UID: \"54e431c7-f38b-4538-a9f6-69d7a8995dce\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tdwnv" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.375139 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1ea01a0b-ae00-469d-a56d-0103ac067d1b-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-z6qdr\" (UID: \"1ea01a0b-ae00-469d-a56d-0103ac067d1b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-z6qdr" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.375160 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hg4w\" (UniqueName: \"kubernetes.io/projected/f4709124-2f04-4d3c-abb7-30a5d0925b6c-kube-api-access-4hg4w\") pod \"control-plane-machine-set-operator-78cbb6b69f-xg89d\" (UID: \"f4709124-2f04-4d3c-abb7-30a5d0925b6c\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xg89d" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.375182 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54e431c7-f38b-4538-a9f6-69d7a8995dce-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tdwnv\" (UID: \"54e431c7-f38b-4538-a9f6-69d7a8995dce\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tdwnv" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.375220 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nf47\" (UniqueName: \"kubernetes.io/projected/40a9c91a-1a6d-438d-9c65-51dc38b34f34-kube-api-access-8nf47\") pod \"controller-manager-879f6c89f-kw4m8\" (UID: \"40a9c91a-1a6d-438d-9c65-51dc38b34f34\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kw4m8" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.375241 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1ea01a0b-ae00-469d-a56d-0103ac067d1b-service-ca-bundle\") pod \"authentication-operator-69f744f599-z6qdr\" (UID: \"1ea01a0b-ae00-469d-a56d-0103ac067d1b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-z6qdr" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.375264 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/01924913-40f0-4502-b151-36079ee0ea8c-audit-dir\") pod \"apiserver-7bbb656c7d-lcqht\" (UID: \"01924913-40f0-4502-b151-36079ee0ea8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lcqht" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.375287 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7zfc\" (UniqueName: \"kubernetes.io/projected/af90c16f-64c2-425f-843e-dd3a96fa9533-kube-api-access-p7zfc\") pod \"openshift-config-operator-7777fb866f-cl5cd\" (UID: \"af90c16f-64c2-425f-843e-dd3a96fa9533\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cl5cd" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.375315 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40a9c91a-1a6d-438d-9c65-51dc38b34f34-serving-cert\") pod \"controller-manager-879f6c89f-kw4m8\" (UID: \"40a9c91a-1a6d-438d-9c65-51dc38b34f34\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kw4m8" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.375335 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/f4709124-2f04-4d3c-abb7-30a5d0925b6c-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-xg89d\" (UID: \"f4709124-2f04-4d3c-abb7-30a5d0925b6c\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xg89d" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.375354 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ea01a0b-ae00-469d-a56d-0103ac067d1b-serving-cert\") pod \"authentication-operator-69f744f599-z6qdr\" (UID: \"1ea01a0b-ae00-469d-a56d-0103ac067d1b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-z6qdr" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.375385 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/fb8fcfca-0933-4a74-85de-859fda0153d6-images\") pod \"machine-api-operator-5694c8668f-52pjn\" (UID: \"fb8fcfca-0933-4a74-85de-859fda0153d6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-52pjn" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.375413 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x99bn\" (UniqueName: \"kubernetes.io/projected/fb8fcfca-0933-4a74-85de-859fda0153d6-kube-api-access-x99bn\") pod \"machine-api-operator-5694c8668f-52pjn\" (UID: \"fb8fcfca-0933-4a74-85de-859fda0153d6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-52pjn" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.375437 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/fb8fcfca-0933-4a74-85de-859fda0153d6-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-52pjn\" (UID: \"fb8fcfca-0933-4a74-85de-859fda0153d6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-52pjn" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.375457 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/af90c16f-64c2-425f-843e-dd3a96fa9533-serving-cert\") pod \"openshift-config-operator-7777fb866f-cl5cd\" (UID: \"af90c16f-64c2-425f-843e-dd3a96fa9533\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cl5cd" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.375477 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/01924913-40f0-4502-b151-36079ee0ea8c-encryption-config\") pod \"apiserver-7bbb656c7d-lcqht\" (UID: \"01924913-40f0-4502-b151-36079ee0ea8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lcqht" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.375494 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ea01a0b-ae00-469d-a56d-0103ac067d1b-config\") pod \"authentication-operator-69f744f599-z6qdr\" (UID: \"1ea01a0b-ae00-469d-a56d-0103ac067d1b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-z6qdr" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.375516 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/af90c16f-64c2-425f-843e-dd3a96fa9533-available-featuregates\") pod \"openshift-config-operator-7777fb866f-cl5cd\" (UID: \"af90c16f-64c2-425f-843e-dd3a96fa9533\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cl5cd" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.375535 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/01924913-40f0-4502-b151-36079ee0ea8c-etcd-client\") pod \"apiserver-7bbb656c7d-lcqht\" (UID: \"01924913-40f0-4502-b151-36079ee0ea8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lcqht" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.375552 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/40a9c91a-1a6d-438d-9c65-51dc38b34f34-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-kw4m8\" (UID: \"40a9c91a-1a6d-438d-9c65-51dc38b34f34\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kw4m8" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.375569 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54e431c7-f38b-4538-a9f6-69d7a8995dce-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tdwnv\" (UID: \"54e431c7-f38b-4538-a9f6-69d7a8995dce\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tdwnv" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.375978 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/01924913-40f0-4502-b151-36079ee0ea8c-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-lcqht\" (UID: \"01924913-40f0-4502-b151-36079ee0ea8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lcqht" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.378170 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/01924913-40f0-4502-b151-36079ee0ea8c-audit-policies\") pod \"apiserver-7bbb656c7d-lcqht\" (UID: \"01924913-40f0-4502-b151-36079ee0ea8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lcqht" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.378907 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/40a9c91a-1a6d-438d-9c65-51dc38b34f34-client-ca\") pod \"controller-manager-879f6c89f-kw4m8\" (UID: \"40a9c91a-1a6d-438d-9c65-51dc38b34f34\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kw4m8" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.380380 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/01924913-40f0-4502-b151-36079ee0ea8c-audit-dir\") pod \"apiserver-7bbb656c7d-lcqht\" (UID: \"01924913-40f0-4502-b151-36079ee0ea8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lcqht" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.381754 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/40a9c91a-1a6d-438d-9c65-51dc38b34f34-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-kw4m8\" (UID: \"40a9c91a-1a6d-438d-9c65-51dc38b34f34\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kw4m8" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.383752 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/01924913-40f0-4502-b151-36079ee0ea8c-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-lcqht\" (UID: \"01924913-40f0-4502-b151-36079ee0ea8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lcqht" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.384444 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb8fcfca-0933-4a74-85de-859fda0153d6-config\") pod \"machine-api-operator-5694c8668f-52pjn\" (UID: \"fb8fcfca-0933-4a74-85de-859fda0153d6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-52pjn" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.385608 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40a9c91a-1a6d-438d-9c65-51dc38b34f34-config\") pod \"controller-manager-879f6c89f-kw4m8\" (UID: \"40a9c91a-1a6d-438d-9c65-51dc38b34f34\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kw4m8" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.388882 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/fb8fcfca-0933-4a74-85de-859fda0153d6-images\") pod \"machine-api-operator-5694c8668f-52pjn\" (UID: \"fb8fcfca-0933-4a74-85de-859fda0153d6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-52pjn" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.391769 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-csmdk"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.392514 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-f5kts"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.393156 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-f5kts" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.393489 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-csmdk" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.395305 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-gvkmx"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.396236 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-gvkmx" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.397663 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40a9c91a-1a6d-438d-9c65-51dc38b34f34-serving-cert\") pod \"controller-manager-879f6c89f-kw4m8\" (UID: \"40a9c91a-1a6d-438d-9c65-51dc38b34f34\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kw4m8" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.398271 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.398379 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.399579 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.399748 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/fb8fcfca-0933-4a74-85de-859fda0153d6-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-52pjn\" (UID: \"fb8fcfca-0933-4a74-85de-859fda0153d6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-52pjn" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.400196 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.405409 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-l9mq8"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.406289 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-l9mq8" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.406334 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-x9tsn"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.406805 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.406925 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-x9tsn" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.408532 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/01924913-40f0-4502-b151-36079ee0ea8c-etcd-client\") pod \"apiserver-7bbb656c7d-lcqht\" (UID: \"01924913-40f0-4502-b151-36079ee0ea8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lcqht" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.408590 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-s5ct4"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.408628 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01924913-40f0-4502-b151-36079ee0ea8c-serving-cert\") pod \"apiserver-7bbb656c7d-lcqht\" (UID: \"01924913-40f0-4502-b151-36079ee0ea8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lcqht" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.408977 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.409217 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.410690 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s5ct4" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.411164 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.412862 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/01924913-40f0-4502-b151-36079ee0ea8c-encryption-config\") pod \"apiserver-7bbb656c7d-lcqht\" (UID: \"01924913-40f0-4502-b151-36079ee0ea8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lcqht" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.413192 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.413463 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.413877 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.414034 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.415118 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.415320 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.415419 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.415532 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.417390 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.417935 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.418082 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-whl8s"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.418510 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.418642 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-lcqht"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.418743 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-whl8s" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.422605 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-kw4m8"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.425585 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.427739 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.429465 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.429593 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.440114 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-5bhp6"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.440170 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-xt48p"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.440189 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-52pjn"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.440205 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-z6qdr"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.441386 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gscbx"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.443404 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409630-lnw2h"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.444356 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-lnw2h" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.446122 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.448832 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.449114 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tdwnv"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.453371 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-gxhcq"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.463183 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.465564 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-q4wmp"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.465626 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-76drh"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.467061 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-csmdk"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.471505 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xrgx6"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.472468 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-sg6bg"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.472663 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xrgx6" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.474219 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ptnr8"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.476219 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-cl5cd"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.478056 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/af90c16f-64c2-425f-843e-dd3a96fa9533-serving-cert\") pod \"openshift-config-operator-7777fb866f-cl5cd\" (UID: \"af90c16f-64c2-425f-843e-dd3a96fa9533\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cl5cd" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.478212 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/af90c16f-64c2-425f-843e-dd3a96fa9533-available-featuregates\") pod \"openshift-config-operator-7777fb866f-cl5cd\" (UID: \"af90c16f-64c2-425f-843e-dd3a96fa9533\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cl5cd" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.478449 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ea01a0b-ae00-469d-a56d-0103ac067d1b-config\") pod \"authentication-operator-69f744f599-z6qdr\" (UID: \"1ea01a0b-ae00-469d-a56d-0103ac067d1b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-z6qdr" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.478581 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54e431c7-f38b-4538-a9f6-69d7a8995dce-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tdwnv\" (UID: \"54e431c7-f38b-4538-a9f6-69d7a8995dce\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tdwnv" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.478694 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjzcq\" (UniqueName: \"kubernetes.io/projected/dc0adee4-65bf-484b-8f55-3c83022cae82-kube-api-access-cjzcq\") pod \"downloads-7954f5f757-76drh\" (UID: \"dc0adee4-65bf-484b-8f55-3c83022cae82\") " pod="openshift-console/downloads-7954f5f757-76drh" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.478821 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swdsl\" (UniqueName: \"kubernetes.io/projected/1ea01a0b-ae00-469d-a56d-0103ac067d1b-kube-api-access-swdsl\") pod \"authentication-operator-69f744f599-z6qdr\" (UID: \"1ea01a0b-ae00-469d-a56d-0103ac067d1b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-z6qdr" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.478921 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/54e431c7-f38b-4538-a9f6-69d7a8995dce-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tdwnv\" (UID: \"54e431c7-f38b-4538-a9f6-69d7a8995dce\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tdwnv" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.479167 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hg4w\" (UniqueName: \"kubernetes.io/projected/f4709124-2f04-4d3c-abb7-30a5d0925b6c-kube-api-access-4hg4w\") pod \"control-plane-machine-set-operator-78cbb6b69f-xg89d\" (UID: \"f4709124-2f04-4d3c-abb7-30a5d0925b6c\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xg89d" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.479260 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/af90c16f-64c2-425f-843e-dd3a96fa9533-available-featuregates\") pod \"openshift-config-operator-7777fb866f-cl5cd\" (UID: \"af90c16f-64c2-425f-843e-dd3a96fa9533\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cl5cd" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.479290 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1ea01a0b-ae00-469d-a56d-0103ac067d1b-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-z6qdr\" (UID: \"1ea01a0b-ae00-469d-a56d-0103ac067d1b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-z6qdr" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.479488 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54e431c7-f38b-4538-a9f6-69d7a8995dce-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tdwnv\" (UID: \"54e431c7-f38b-4538-a9f6-69d7a8995dce\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tdwnv" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.479952 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1ea01a0b-ae00-469d-a56d-0103ac067d1b-service-ca-bundle\") pod \"authentication-operator-69f744f599-z6qdr\" (UID: \"1ea01a0b-ae00-469d-a56d-0103ac067d1b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-z6qdr" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.480340 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ea01a0b-ae00-469d-a56d-0103ac067d1b-config\") pod \"authentication-operator-69f744f599-z6qdr\" (UID: \"1ea01a0b-ae00-469d-a56d-0103ac067d1b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-z6qdr" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.480875 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1ea01a0b-ae00-469d-a56d-0103ac067d1b-service-ca-bundle\") pod \"authentication-operator-69f744f599-z6qdr\" (UID: \"1ea01a0b-ae00-469d-a56d-0103ac067d1b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-z6qdr" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.481129 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54e431c7-f38b-4538-a9f6-69d7a8995dce-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tdwnv\" (UID: \"54e431c7-f38b-4538-a9f6-69d7a8995dce\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tdwnv" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.481289 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7zfc\" (UniqueName: \"kubernetes.io/projected/af90c16f-64c2-425f-843e-dd3a96fa9533-kube-api-access-p7zfc\") pod \"openshift-config-operator-7777fb866f-cl5cd\" (UID: \"af90c16f-64c2-425f-843e-dd3a96fa9533\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cl5cd" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.481320 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/f4709124-2f04-4d3c-abb7-30a5d0925b6c-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-xg89d\" (UID: \"f4709124-2f04-4d3c-abb7-30a5d0925b6c\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xg89d" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.481616 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ea01a0b-ae00-469d-a56d-0103ac067d1b-serving-cert\") pod \"authentication-operator-69f744f599-z6qdr\" (UID: \"1ea01a0b-ae00-469d-a56d-0103ac067d1b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-z6qdr" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.481971 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.482189 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f5rc8"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.482554 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1ea01a0b-ae00-469d-a56d-0103ac067d1b-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-z6qdr\" (UID: \"1ea01a0b-ae00-469d-a56d-0103ac067d1b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-z6qdr" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.482980 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f5rc8" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.483123 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54e431c7-f38b-4538-a9f6-69d7a8995dce-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tdwnv\" (UID: \"54e431c7-f38b-4538-a9f6-69d7a8995dce\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tdwnv" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.484221 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7dcf8"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.484979 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-nd5kp"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.485250 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/f4709124-2f04-4d3c-abb7-30a5d0925b6c-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-xg89d\" (UID: \"f4709124-2f04-4d3c-abb7-30a5d0925b6c\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xg89d" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.485451 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ea01a0b-ae00-469d-a56d-0103ac067d1b-serving-cert\") pod \"authentication-operator-69f744f599-z6qdr\" (UID: \"1ea01a0b-ae00-469d-a56d-0103ac067d1b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-z6qdr" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.485899 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cjn5c"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.487405 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xfb2s"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.488694 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-w8xn2"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.489861 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-f5kts"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.496989 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xg89d"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.499043 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-pp9bw"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.499847 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-pp9bw" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.501571 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.501737 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-l9mq8"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.502980 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-292nl"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.504171 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-qvzhc"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.507294 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7hstm"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.508391 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xdpgx"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.509598 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mvczl"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.510614 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mvczl" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.510859 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-crgj6"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.511761 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-crgj6" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.513299 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-vq52n"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.513934 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-vq52n" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.516242 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f5rc8"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.517660 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xrgx6"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.519038 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-gvkmx"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.519956 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-pp9bw"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.521247 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-s5ct4"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.523195 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-whl8s"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.523417 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-x9tsn"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.524599 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.524870 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mvczl"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.527251 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-9nmnq"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.528582 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-9nmnq" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.530773 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-crgj6"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.531867 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409630-lnw2h"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.532935 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-9nmnq"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.534026 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-pwc6t"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.534714 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-pwc6t" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.538751 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-pwc6t"] Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.542577 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.561753 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.581388 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.602752 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.621489 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.641861 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.661769 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.697675 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.702111 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.722557 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.734347 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/af90c16f-64c2-425f-843e-dd3a96fa9533-serving-cert\") pod \"openshift-config-operator-7777fb866f-cl5cd\" (UID: \"af90c16f-64c2-425f-843e-dd3a96fa9533\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cl5cd" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.742854 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.762190 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.782740 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.802589 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.822178 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.863194 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.882412 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.913129 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.923306 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.943333 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.964196 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 01 08:42:38 crc kubenswrapper[4873]: I1201 08:42:38.983437 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.003483 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.023240 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.042123 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.062360 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.082651 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.103006 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.123817 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.143103 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.162979 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.183647 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.227894 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x99bn\" (UniqueName: \"kubernetes.io/projected/fb8fcfca-0933-4a74-85de-859fda0153d6-kube-api-access-x99bn\") pod \"machine-api-operator-5694c8668f-52pjn\" (UID: \"fb8fcfca-0933-4a74-85de-859fda0153d6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-52pjn" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.249283 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nf47\" (UniqueName: \"kubernetes.io/projected/40a9c91a-1a6d-438d-9c65-51dc38b34f34-kube-api-access-8nf47\") pod \"controller-manager-879f6c89f-kw4m8\" (UID: \"40a9c91a-1a6d-438d-9c65-51dc38b34f34\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kw4m8" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.263089 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.275998 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdzlb\" (UniqueName: \"kubernetes.io/projected/01924913-40f0-4502-b151-36079ee0ea8c-kube-api-access-tdzlb\") pod \"apiserver-7bbb656c7d-lcqht\" (UID: \"01924913-40f0-4502-b151-36079ee0ea8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lcqht" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.283396 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.303594 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.323798 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.343215 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.363349 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.383437 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.400727 4873 request.go:700] Waited for 1.006853251s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver-operator/configmaps?fieldSelector=metadata.name%3Dkube-apiserver-operator-config&limit=500&resourceVersion=0 Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.402897 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.423600 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.442791 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lcqht" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.453159 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.462679 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.476127 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-52pjn" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.484153 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.499761 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-kw4m8" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.504233 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.524236 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.542517 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.563345 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.583190 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.603187 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.624227 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.642598 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.663585 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.683132 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.702128 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.723426 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-lcqht"] Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.725912 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.758162 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-52pjn"] Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.759164 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-kw4m8"] Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.762410 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 01 08:42:39 crc kubenswrapper[4873]: W1201 08:42:39.770442 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod40a9c91a_1a6d_438d_9c65_51dc38b34f34.slice/crio-80c19d1e61c087c6b2acfbee6dd4fbbbe839dd042bde472e13444d35c71f2c0a WatchSource:0}: Error finding container 80c19d1e61c087c6b2acfbee6dd4fbbbe839dd042bde472e13444d35c71f2c0a: Status 404 returned error can't find the container with id 80c19d1e61c087c6b2acfbee6dd4fbbbe839dd042bde472e13444d35c71f2c0a Dec 01 08:42:39 crc kubenswrapper[4873]: W1201 08:42:39.770750 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfb8fcfca_0933_4a74_85de_859fda0153d6.slice/crio-51466f6de2318b2fa13b593cf3551875488136b742d850f4de33c9e01af9e9a3 WatchSource:0}: Error finding container 51466f6de2318b2fa13b593cf3551875488136b742d850f4de33c9e01af9e9a3: Status 404 returned error can't find the container with id 51466f6de2318b2fa13b593cf3551875488136b742d850f4de33c9e01af9e9a3 Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.781666 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.802897 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.823359 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.843650 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.861948 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.882141 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.902883 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.944843 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/54e431c7-f38b-4538-a9f6-69d7a8995dce-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tdwnv\" (UID: \"54e431c7-f38b-4538-a9f6-69d7a8995dce\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tdwnv" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.967691 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjzcq\" (UniqueName: \"kubernetes.io/projected/dc0adee4-65bf-484b-8f55-3c83022cae82-kube-api-access-cjzcq\") pod \"downloads-7954f5f757-76drh\" (UID: \"dc0adee4-65bf-484b-8f55-3c83022cae82\") " pod="openshift-console/downloads-7954f5f757-76drh" Dec 01 08:42:39 crc kubenswrapper[4873]: I1201 08:42:39.982230 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swdsl\" (UniqueName: \"kubernetes.io/projected/1ea01a0b-ae00-469d-a56d-0103ac067d1b-kube-api-access-swdsl\") pod \"authentication-operator-69f744f599-z6qdr\" (UID: \"1ea01a0b-ae00-469d-a56d-0103ac067d1b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-z6qdr" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.003099 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hg4w\" (UniqueName: \"kubernetes.io/projected/f4709124-2f04-4d3c-abb7-30a5d0925b6c-kube-api-access-4hg4w\") pod \"control-plane-machine-set-operator-78cbb6b69f-xg89d\" (UID: \"f4709124-2f04-4d3c-abb7-30a5d0925b6c\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xg89d" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.015831 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7zfc\" (UniqueName: \"kubernetes.io/projected/af90c16f-64c2-425f-843e-dd3a96fa9533-kube-api-access-p7zfc\") pod \"openshift-config-operator-7777fb866f-cl5cd\" (UID: \"af90c16f-64c2-425f-843e-dd3a96fa9533\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cl5cd" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.021880 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.044798 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-76drh" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.047040 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.050865 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-z6qdr" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.062658 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.065523 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xg89d" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.082583 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.086896 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cl5cd" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.103522 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.123707 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.143916 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.163751 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.183268 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.203798 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.223193 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.240138 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tdwnv" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.242381 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.264230 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.282577 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.301809 4873 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.324089 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.350593 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.363838 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.385378 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.405559 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.420504 4873 request.go:700] Waited for 1.576957643s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/persistentvolumes/pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.426819 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-cl5cd"] Dec 01 08:42:40 crc kubenswrapper[4873]: W1201 08:42:40.439832 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaf90c16f_64c2_425f_843e_dd3a96fa9533.slice/crio-3a6e88a5d58010bd441de3dd89de7ff6aeed5ef7f4fcf365a7d42759f4d58203 WatchSource:0}: Error finding container 3a6e88a5d58010bd441de3dd89de7ff6aeed5ef7f4fcf365a7d42759f4d58203: Status 404 returned error can't find the container with id 3a6e88a5d58010bd441de3dd89de7ff6aeed5ef7f4fcf365a7d42759f4d58203 Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.467609 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tdwnv"] Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.476136 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cl5cd" event={"ID":"af90c16f-64c2-425f-843e-dd3a96fa9533","Type":"ContainerStarted","Data":"3a6e88a5d58010bd441de3dd89de7ff6aeed5ef7f4fcf365a7d42759f4d58203"} Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.477512 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-52pjn" event={"ID":"fb8fcfca-0933-4a74-85de-859fda0153d6","Type":"ContainerStarted","Data":"6f300deeacaf0ac420ef14fdd65fd2d9675ea27c514852edacc1b963dfa5ec6b"} Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.477566 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-52pjn" event={"ID":"fb8fcfca-0933-4a74-85de-859fda0153d6","Type":"ContainerStarted","Data":"51466f6de2318b2fa13b593cf3551875488136b742d850f4de33c9e01af9e9a3"} Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.478915 4873 generic.go:334] "Generic (PLEG): container finished" podID="01924913-40f0-4502-b151-36079ee0ea8c" containerID="26bf31b37193341dc872fcf7b211bfe2155ca12aacdac1eaf1c7226cda350439" exitCode=0 Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.478973 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lcqht" event={"ID":"01924913-40f0-4502-b151-36079ee0ea8c","Type":"ContainerDied","Data":"26bf31b37193341dc872fcf7b211bfe2155ca12aacdac1eaf1c7226cda350439"} Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.479118 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lcqht" event={"ID":"01924913-40f0-4502-b151-36079ee0ea8c","Type":"ContainerStarted","Data":"6e8a1723b6700199814157a7ad61cacd8e8657e40780de46367e31b3d63f2d68"} Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.484681 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-kw4m8" event={"ID":"40a9c91a-1a6d-438d-9c65-51dc38b34f34","Type":"ContainerStarted","Data":"e0c3f9fbfe354bd1a18ec89ef06c19691b818ad2eaac663b4713aebd080776a7"} Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.484741 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-kw4m8" event={"ID":"40a9c91a-1a6d-438d-9c65-51dc38b34f34","Type":"ContainerStarted","Data":"80c19d1e61c087c6b2acfbee6dd4fbbbe839dd042bde472e13444d35c71f2c0a"} Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.485040 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-kw4m8" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.496417 4873 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-kw4m8 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.496474 4873 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-kw4m8" podUID="40a9c91a-1a6d-438d-9c65-51dc38b34f34" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.509946 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/16511918-7dea-4667-9119-07f0da1b9a84-serving-cert\") pod \"console-operator-58897d9998-5bhp6\" (UID: \"16511918-7dea-4667-9119-07f0da1b9a84\") " pod="openshift-console-operator/console-operator-58897d9998-5bhp6" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.509990 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/757ee90a-18e3-4ccb-b03c-43e90bf8cb34-etcd-ca\") pod \"etcd-operator-b45778765-xt48p\" (UID: \"757ee90a-18e3-4ccb-b03c-43e90bf8cb34\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xt48p" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.510031 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0d0699e2-c1b3-48a3-8d17-5d89424b4bb0-auth-proxy-config\") pod \"machine-approver-56656f9798-wq2zz\" (UID: \"0d0699e2-c1b3-48a3-8d17-5d89424b4bb0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wq2zz" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.510052 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-sg6bg\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.510074 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/952ae5ba-156f-41a7-950a-1783e2e69f94-registry-tls\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.510099 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2404e1f8-203a-4055-8f66-21d8867e7a92-audit-policies\") pod \"oauth-openshift-558db77b4-sg6bg\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.510117 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-sg6bg\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.510169 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-sg6bg\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.510193 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/0d0699e2-c1b3-48a3-8d17-5d89424b4bb0-machine-approver-tls\") pod \"machine-approver-56656f9798-wq2zz\" (UID: \"0d0699e2-c1b3-48a3-8d17-5d89424b4bb0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wq2zz" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.510307 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-sg6bg\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.510325 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-sg6bg\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.510381 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/16511918-7dea-4667-9119-07f0da1b9a84-trusted-ca\") pod \"console-operator-58897d9998-5bhp6\" (UID: \"16511918-7dea-4667-9119-07f0da1b9a84\") " pod="openshift-console-operator/console-operator-58897d9998-5bhp6" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.510398 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25wjr\" (UniqueName: \"kubernetes.io/projected/4af31a76-3090-497d-b39f-965c27a27a1c-kube-api-access-25wjr\") pod \"machine-config-controller-84d6567774-gxhcq\" (UID: \"4af31a76-3090-497d-b39f-965c27a27a1c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gxhcq" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.510460 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ae767d9a-3df4-413a-bc25-8fc819c7088e-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-xfb2s\" (UID: \"ae767d9a-3df4-413a-bc25-8fc819c7088e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xfb2s" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.510493 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/952ae5ba-156f-41a7-950a-1783e2e69f94-trusted-ca\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.510551 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwvvt\" (UniqueName: \"kubernetes.io/projected/10d7dabb-aa6f-4f2b-acf9-b02dc369d8be-kube-api-access-rwvvt\") pod \"cluster-samples-operator-665b6dd947-nd5kp\" (UID: \"10d7dabb-aa6f-4f2b-acf9-b02dc369d8be\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-nd5kp" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.510571 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-sg6bg\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.510628 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9491cb1f-acf6-438b-8175-11737d7bd245-oauth-serving-cert\") pod \"console-f9d7485db-292nl\" (UID: \"9491cb1f-acf6-438b-8175-11737d7bd245\") " pod="openshift-console/console-f9d7485db-292nl" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.510648 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgbxj\" (UniqueName: \"kubernetes.io/projected/16511918-7dea-4667-9119-07f0da1b9a84-kube-api-access-wgbxj\") pod \"console-operator-58897d9998-5bhp6\" (UID: \"16511918-7dea-4667-9119-07f0da1b9a84\") " pod="openshift-console-operator/console-operator-58897d9998-5bhp6" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.510667 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d0c705dd-9608-4b5e-8c91-2511d8427419-serving-cert\") pod \"apiserver-76f77b778f-w8xn2\" (UID: \"d0c705dd-9608-4b5e-8c91-2511d8427419\") " pod="openshift-apiserver/apiserver-76f77b778f-w8xn2" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.510684 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d0699e2-c1b3-48a3-8d17-5d89424b4bb0-config\") pod \"machine-approver-56656f9798-wq2zz\" (UID: \"0d0699e2-c1b3-48a3-8d17-5d89424b4bb0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wq2zz" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.510713 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4af31a76-3090-497d-b39f-965c27a27a1c-proxy-tls\") pod \"machine-config-controller-84d6567774-gxhcq\" (UID: \"4af31a76-3090-497d-b39f-965c27a27a1c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gxhcq" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.510736 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9491cb1f-acf6-438b-8175-11737d7bd245-console-oauth-config\") pod \"console-f9d7485db-292nl\" (UID: \"9491cb1f-acf6-438b-8175-11737d7bd245\") " pod="openshift-console/console-f9d7485db-292nl" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.510770 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/d0c705dd-9608-4b5e-8c91-2511d8427419-image-import-ca\") pod \"apiserver-76f77b778f-w8xn2\" (UID: \"d0c705dd-9608-4b5e-8c91-2511d8427419\") " pod="openshift-apiserver/apiserver-76f77b778f-w8xn2" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.510789 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xv9q\" (UniqueName: \"kubernetes.io/projected/2404e1f8-203a-4055-8f66-21d8867e7a92-kube-api-access-4xv9q\") pod \"oauth-openshift-558db77b4-sg6bg\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.510820 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.510838 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/952ae5ba-156f-41a7-950a-1783e2e69f94-bound-sa-token\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.510858 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z78ht\" (UniqueName: \"kubernetes.io/projected/17e954cb-2d58-4616-bd5c-e813644d411d-kube-api-access-z78ht\") pod \"multus-admission-controller-857f4d67dd-qvzhc\" (UID: \"17e954cb-2d58-4616-bd5c-e813644d411d\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-qvzhc" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.510890 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/952ae5ba-156f-41a7-950a-1783e2e69f94-ca-trust-extracted\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.510906 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/952ae5ba-156f-41a7-950a-1783e2e69f94-registry-certificates\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.510924 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d0c705dd-9608-4b5e-8c91-2511d8427419-node-pullsecrets\") pod \"apiserver-76f77b778f-w8xn2\" (UID: \"d0c705dd-9608-4b5e-8c91-2511d8427419\") " pod="openshift-apiserver/apiserver-76f77b778f-w8xn2" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.510942 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d0c705dd-9608-4b5e-8c91-2511d8427419-encryption-config\") pod \"apiserver-76f77b778f-w8xn2\" (UID: \"d0c705dd-9608-4b5e-8c91-2511d8427419\") " pod="openshift-apiserver/apiserver-76f77b778f-w8xn2" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.510958 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-sg6bg\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.510976 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5-serving-cert\") pod \"route-controller-manager-6576b87f9c-q4wmp\" (UID: \"83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q4wmp" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.511008 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/645ef95b-f94d-4ba4-8f10-fbe046a9d9a7-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-7dcf8\" (UID: \"645ef95b-f94d-4ba4-8f10-fbe046a9d9a7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7dcf8" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.511040 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-sg6bg\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.511058 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d0c705dd-9608-4b5e-8c91-2511d8427419-trusted-ca-bundle\") pod \"apiserver-76f77b778f-w8xn2\" (UID: \"d0c705dd-9608-4b5e-8c91-2511d8427419\") " pod="openshift-apiserver/apiserver-76f77b778f-w8xn2" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.511086 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlc6g\" (UniqueName: \"kubernetes.io/projected/952ae5ba-156f-41a7-950a-1783e2e69f94-kube-api-access-qlc6g\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.511103 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwdlq\" (UniqueName: \"kubernetes.io/projected/757ee90a-18e3-4ccb-b03c-43e90bf8cb34-kube-api-access-gwdlq\") pod \"etcd-operator-b45778765-xt48p\" (UID: \"757ee90a-18e3-4ccb-b03c-43e90bf8cb34\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xt48p" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.511118 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2404e1f8-203a-4055-8f66-21d8867e7a92-audit-dir\") pod \"oauth-openshift-558db77b4-sg6bg\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.511134 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-sg6bg\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.511153 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-sg6bg\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.511173 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qwvk\" (UniqueName: \"kubernetes.io/projected/83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5-kube-api-access-4qwvk\") pod \"route-controller-manager-6576b87f9c-q4wmp\" (UID: \"83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q4wmp" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.511191 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/645ef95b-f94d-4ba4-8f10-fbe046a9d9a7-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-7dcf8\" (UID: \"645ef95b-f94d-4ba4-8f10-fbe046a9d9a7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7dcf8" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.511208 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9491cb1f-acf6-438b-8175-11737d7bd245-service-ca\") pod \"console-f9d7485db-292nl\" (UID: \"9491cb1f-acf6-438b-8175-11737d7bd245\") " pod="openshift-console/console-f9d7485db-292nl" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.511224 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9491cb1f-acf6-438b-8175-11737d7bd245-trusted-ca-bundle\") pod \"console-f9d7485db-292nl\" (UID: \"9491cb1f-acf6-438b-8175-11737d7bd245\") " pod="openshift-console/console-f9d7485db-292nl" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.511252 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c92cq\" (UniqueName: \"kubernetes.io/projected/ae767d9a-3df4-413a-bc25-8fc819c7088e-kube-api-access-c92cq\") pod \"openshift-apiserver-operator-796bbdcf4f-xfb2s\" (UID: \"ae767d9a-3df4-413a-bc25-8fc819c7088e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xfb2s" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.511274 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fd09645-8bbe-458c-8d9b-b4c2875fd963-config\") pod \"kube-controller-manager-operator-78b949d7b-gscbx\" (UID: \"4fd09645-8bbe-458c-8d9b-b4c2875fd963\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gscbx" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.511295 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f2208bec-45d1-45f0-8700-1218abfa7642-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-ptnr8\" (UID: \"f2208bec-45d1-45f0-8700-1218abfa7642\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ptnr8" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.511313 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7q6cm\" (UniqueName: \"kubernetes.io/projected/f2208bec-45d1-45f0-8700-1218abfa7642-kube-api-access-7q6cm\") pod \"cluster-image-registry-operator-dc59b4c8b-ptnr8\" (UID: \"f2208bec-45d1-45f0-8700-1218abfa7642\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ptnr8" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.511329 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4fd09645-8bbe-458c-8d9b-b4c2875fd963-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-gscbx\" (UID: \"4fd09645-8bbe-458c-8d9b-b4c2875fd963\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gscbx" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.511345 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9491cb1f-acf6-438b-8175-11737d7bd245-console-config\") pod \"console-f9d7485db-292nl\" (UID: \"9491cb1f-acf6-438b-8175-11737d7bd245\") " pod="openshift-console/console-f9d7485db-292nl" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.511385 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d0c705dd-9608-4b5e-8c91-2511d8427419-etcd-serving-ca\") pod \"apiserver-76f77b778f-w8xn2\" (UID: \"d0c705dd-9608-4b5e-8c91-2511d8427419\") " pod="openshift-apiserver/apiserver-76f77b778f-w8xn2" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.511414 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d0c705dd-9608-4b5e-8c91-2511d8427419-audit-dir\") pod \"apiserver-76f77b778f-w8xn2\" (UID: \"d0c705dd-9608-4b5e-8c91-2511d8427419\") " pod="openshift-apiserver/apiserver-76f77b778f-w8xn2" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.511432 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/10d7dabb-aa6f-4f2b-acf9-b02dc369d8be-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-nd5kp\" (UID: \"10d7dabb-aa6f-4f2b-acf9-b02dc369d8be\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-nd5kp" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.511451 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f2208bec-45d1-45f0-8700-1218abfa7642-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-ptnr8\" (UID: \"f2208bec-45d1-45f0-8700-1218abfa7642\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ptnr8" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.511469 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-sg6bg\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.511486 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5-config\") pod \"route-controller-manager-6576b87f9c-q4wmp\" (UID: \"83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q4wmp" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.511510 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/757ee90a-18e3-4ccb-b03c-43e90bf8cb34-etcd-service-ca\") pod \"etcd-operator-b45778765-xt48p\" (UID: \"757ee90a-18e3-4ccb-b03c-43e90bf8cb34\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xt48p" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.511528 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9th6b\" (UniqueName: \"kubernetes.io/projected/0d0699e2-c1b3-48a3-8d17-5d89424b4bb0-kube-api-access-9th6b\") pod \"machine-approver-56656f9798-wq2zz\" (UID: \"0d0699e2-c1b3-48a3-8d17-5d89424b4bb0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wq2zz" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.511555 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0c705dd-9608-4b5e-8c91-2511d8427419-config\") pod \"apiserver-76f77b778f-w8xn2\" (UID: \"d0c705dd-9608-4b5e-8c91-2511d8427419\") " pod="openshift-apiserver/apiserver-76f77b778f-w8xn2" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.511572 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f2208bec-45d1-45f0-8700-1218abfa7642-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-ptnr8\" (UID: \"f2208bec-45d1-45f0-8700-1218abfa7642\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ptnr8" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.511596 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae767d9a-3df4-413a-bc25-8fc819c7088e-config\") pod \"openshift-apiserver-operator-796bbdcf4f-xfb2s\" (UID: \"ae767d9a-3df4-413a-bc25-8fc819c7088e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xfb2s" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.511620 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6h25\" (UniqueName: \"kubernetes.io/projected/645ef95b-f94d-4ba4-8f10-fbe046a9d9a7-kube-api-access-z6h25\") pod \"openshift-controller-manager-operator-756b6f6bc6-7dcf8\" (UID: \"645ef95b-f94d-4ba4-8f10-fbe046a9d9a7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7dcf8" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.511638 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/17e954cb-2d58-4616-bd5c-e813644d411d-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-qvzhc\" (UID: \"17e954cb-2d58-4616-bd5c-e813644d411d\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-qvzhc" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.511657 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4fd09645-8bbe-458c-8d9b-b4c2875fd963-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-gscbx\" (UID: \"4fd09645-8bbe-458c-8d9b-b4c2875fd963\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gscbx" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.511673 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4af31a76-3090-497d-b39f-965c27a27a1c-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-gxhcq\" (UID: \"4af31a76-3090-497d-b39f-965c27a27a1c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gxhcq" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.511689 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5-client-ca\") pod \"route-controller-manager-6576b87f9c-q4wmp\" (UID: \"83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q4wmp" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.511711 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/d0c705dd-9608-4b5e-8c91-2511d8427419-audit\") pod \"apiserver-76f77b778f-w8xn2\" (UID: \"d0c705dd-9608-4b5e-8c91-2511d8427419\") " pod="openshift-apiserver/apiserver-76f77b778f-w8xn2" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.511732 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmww7\" (UniqueName: \"kubernetes.io/projected/d0c705dd-9608-4b5e-8c91-2511d8427419-kube-api-access-cmww7\") pod \"apiserver-76f77b778f-w8xn2\" (UID: \"d0c705dd-9608-4b5e-8c91-2511d8427419\") " pod="openshift-apiserver/apiserver-76f77b778f-w8xn2" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.511753 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/757ee90a-18e3-4ccb-b03c-43e90bf8cb34-config\") pod \"etcd-operator-b45778765-xt48p\" (UID: \"757ee90a-18e3-4ccb-b03c-43e90bf8cb34\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xt48p" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.511777 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/757ee90a-18e3-4ccb-b03c-43e90bf8cb34-serving-cert\") pod \"etcd-operator-b45778765-xt48p\" (UID: \"757ee90a-18e3-4ccb-b03c-43e90bf8cb34\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xt48p" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.511802 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16511918-7dea-4667-9119-07f0da1b9a84-config\") pod \"console-operator-58897d9998-5bhp6\" (UID: \"16511918-7dea-4667-9119-07f0da1b9a84\") " pod="openshift-console-operator/console-operator-58897d9998-5bhp6" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.511847 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d0c705dd-9608-4b5e-8c91-2511d8427419-etcd-client\") pod \"apiserver-76f77b778f-w8xn2\" (UID: \"d0c705dd-9608-4b5e-8c91-2511d8427419\") " pod="openshift-apiserver/apiserver-76f77b778f-w8xn2" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.511868 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/757ee90a-18e3-4ccb-b03c-43e90bf8cb34-etcd-client\") pod \"etcd-operator-b45778765-xt48p\" (UID: \"757ee90a-18e3-4ccb-b03c-43e90bf8cb34\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xt48p" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.511886 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9491cb1f-acf6-438b-8175-11737d7bd245-console-serving-cert\") pod \"console-f9d7485db-292nl\" (UID: \"9491cb1f-acf6-438b-8175-11737d7bd245\") " pod="openshift-console/console-f9d7485db-292nl" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.511904 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wggrb\" (UniqueName: \"kubernetes.io/projected/9491cb1f-acf6-438b-8175-11737d7bd245-kube-api-access-wggrb\") pod \"console-f9d7485db-292nl\" (UID: \"9491cb1f-acf6-438b-8175-11737d7bd245\") " pod="openshift-console/console-f9d7485db-292nl" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.511923 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/952ae5ba-156f-41a7-950a-1783e2e69f94-installation-pull-secrets\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:40 crc kubenswrapper[4873]: E1201 08:42:40.513118 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:41.013102861 +0000 UTC m=+136.915211400 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.533129 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-z6qdr"] Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.534400 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xg89d"] Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.537281 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-76drh"] Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.614534 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.614900 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/812f722d-c159-40bb-8901-bac36d182280-config\") pod \"service-ca-operator-777779d784-x9tsn\" (UID: \"812f722d-c159-40bb-8901-bac36d182280\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-x9tsn" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.614939 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8xhk\" (UniqueName: \"kubernetes.io/projected/e3e97e3b-f082-4093-b183-fcd5ac857c07-kube-api-access-n8xhk\") pod \"olm-operator-6b444d44fb-xrgx6\" (UID: \"e3e97e3b-f082-4093-b183-fcd5ac857c07\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xrgx6" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.614986 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f2208bec-45d1-45f0-8700-1218abfa7642-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-ptnr8\" (UID: \"f2208bec-45d1-45f0-8700-1218abfa7642\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ptnr8" Dec 01 08:42:40 crc kubenswrapper[4873]: E1201 08:42:40.615123 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:41.115003073 +0000 UTC m=+137.017111612 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.615160 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9491cb1f-acf6-438b-8175-11737d7bd245-console-config\") pod \"console-f9d7485db-292nl\" (UID: \"9491cb1f-acf6-438b-8175-11737d7bd245\") " pod="openshift-console/console-f9d7485db-292nl" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.615191 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e3e97e3b-f082-4093-b183-fcd5ac857c07-srv-cert\") pod \"olm-operator-6b444d44fb-xrgx6\" (UID: \"e3e97e3b-f082-4093-b183-fcd5ac857c07\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xrgx6" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.615240 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f99594fb-adb1-469b-960f-2a56dc3427bd-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-csmdk\" (UID: \"f99594fb-adb1-469b-960f-2a56dc3427bd\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-csmdk" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.615313 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25z7h\" (UniqueName: \"kubernetes.io/projected/dfbcf0e0-e5a1-41e0-813c-008f355a1f0a-kube-api-access-25z7h\") pod \"migrator-59844c95c7-l9mq8\" (UID: \"dfbcf0e0-e5a1-41e0-813c-008f355a1f0a\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-l9mq8" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.615378 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/10d7dabb-aa6f-4f2b-acf9-b02dc369d8be-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-nd5kp\" (UID: \"10d7dabb-aa6f-4f2b-acf9-b02dc369d8be\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-nd5kp" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.615405 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb6d3b33-6120-40aa-8c4d-01f186d1ab17-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-xdpgx\" (UID: \"fb6d3b33-6120-40aa-8c4d-01f186d1ab17\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xdpgx" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.615445 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e3e97e3b-f082-4093-b183-fcd5ac857c07-profile-collector-cert\") pod \"olm-operator-6b444d44fb-xrgx6\" (UID: \"e3e97e3b-f082-4093-b183-fcd5ac857c07\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xrgx6" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.615493 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2ea695f7-4e9e-4b12-8e0c-22431291576d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7hstm\" (UID: \"2ea695f7-4e9e-4b12-8e0c-22431291576d\") " pod="openshift-marketplace/marketplace-operator-79b997595-7hstm" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.615555 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f2208bec-45d1-45f0-8700-1218abfa7642-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-ptnr8\" (UID: \"f2208bec-45d1-45f0-8700-1218abfa7642\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ptnr8" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.615640 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrqmg\" (UniqueName: \"kubernetes.io/projected/12ddf317-c034-4154-9966-c8b502ec0d6c-kube-api-access-nrqmg\") pod \"service-ca-9c57cc56f-pp9bw\" (UID: \"12ddf317-c034-4154-9966-c8b502ec0d6c\") " pod="openshift-service-ca/service-ca-9c57cc56f-pp9bw" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.615693 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cp6l\" (UniqueName: \"kubernetes.io/projected/9a6d7d20-0eea-42ae-92e0-91b1b555bf6f-kube-api-access-8cp6l\") pod \"dns-default-crgj6\" (UID: \"9a6d7d20-0eea-42ae-92e0-91b1b555bf6f\") " pod="openshift-dns/dns-default-crgj6" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.615720 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3dc1fc6b-9c79-4a78-af19-d49195bc94be-srv-cert\") pod \"catalog-operator-68c6474976-whl8s\" (UID: \"3dc1fc6b-9c79-4a78-af19-d49195bc94be\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-whl8s" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.615787 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae767d9a-3df4-413a-bc25-8fc819c7088e-config\") pod \"openshift-apiserver-operator-796bbdcf4f-xfb2s\" (UID: \"ae767d9a-3df4-413a-bc25-8fc819c7088e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xfb2s" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.615814 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7aee65fe-0d01-4db7-95ee-131e6ead9071-images\") pod \"machine-config-operator-74547568cd-s5ct4\" (UID: \"7aee65fe-0d01-4db7-95ee-131e6ead9071\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s5ct4" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.616203 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9491cb1f-acf6-438b-8175-11737d7bd245-console-config\") pod \"console-f9d7485db-292nl\" (UID: \"9491cb1f-acf6-438b-8175-11737d7bd245\") " pod="openshift-console/console-f9d7485db-292nl" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.616514 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fb6d3b33-6120-40aa-8c4d-01f186d1ab17-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-xdpgx\" (UID: \"fb6d3b33-6120-40aa-8c4d-01f186d1ab17\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xdpgx" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.616924 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae767d9a-3df4-413a-bc25-8fc819c7088e-config\") pod \"openshift-apiserver-operator-796bbdcf4f-xfb2s\" (UID: \"ae767d9a-3df4-413a-bc25-8fc819c7088e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xfb2s" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.617269 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4fd09645-8bbe-458c-8d9b-b4c2875fd963-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-gscbx\" (UID: \"4fd09645-8bbe-458c-8d9b-b4c2875fd963\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gscbx" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.617325 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/2a86c401-c7dd-4b95-b267-8c58e5728e6a-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-mvczl\" (UID: \"2a86c401-c7dd-4b95-b267-8c58e5728e6a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mvczl" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.617356 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/d0c705dd-9608-4b5e-8c91-2511d8427419-audit\") pod \"apiserver-76f77b778f-w8xn2\" (UID: \"d0c705dd-9608-4b5e-8c91-2511d8427419\") " pod="openshift-apiserver/apiserver-76f77b778f-w8xn2" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.617710 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/757ee90a-18e3-4ccb-b03c-43e90bf8cb34-config\") pod \"etcd-operator-b45778765-xt48p\" (UID: \"757ee90a-18e3-4ccb-b03c-43e90bf8cb34\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xt48p" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.617752 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5-client-ca\") pod \"route-controller-manager-6576b87f9c-q4wmp\" (UID: \"83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q4wmp" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.617780 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/04b9f090-d8f6-4122-bbff-af681fcc4f42-mountpoint-dir\") pod \"csi-hostpathplugin-9nmnq\" (UID: \"04b9f090-d8f6-4122-bbff-af681fcc4f42\") " pod="hostpath-provisioner/csi-hostpathplugin-9nmnq" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.618424 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/757ee90a-18e3-4ccb-b03c-43e90bf8cb34-serving-cert\") pod \"etcd-operator-b45778765-xt48p\" (UID: \"757ee90a-18e3-4ccb-b03c-43e90bf8cb34\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xt48p" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.618475 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7hln\" (UniqueName: \"kubernetes.io/projected/812f722d-c159-40bb-8901-bac36d182280-kube-api-access-w7hln\") pod \"service-ca-operator-777779d784-x9tsn\" (UID: \"812f722d-c159-40bb-8901-bac36d182280\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-x9tsn" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.618497 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkwmv\" (UniqueName: \"kubernetes.io/projected/3dc1fc6b-9c79-4a78-af19-d49195bc94be-kube-api-access-mkwmv\") pod \"catalog-operator-68c6474976-whl8s\" (UID: \"3dc1fc6b-9c79-4a78-af19-d49195bc94be\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-whl8s" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.618524 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7n4g\" (UniqueName: \"kubernetes.io/projected/0726d810-1988-4eb3-9077-e8892670dc8c-kube-api-access-z7n4g\") pod \"ingress-operator-5b745b69d9-gvkmx\" (UID: \"0726d810-1988-4eb3-9077-e8892670dc8c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-gvkmx" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.618575 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d0c705dd-9608-4b5e-8c91-2511d8427419-etcd-client\") pod \"apiserver-76f77b778f-w8xn2\" (UID: \"d0c705dd-9608-4b5e-8c91-2511d8427419\") " pod="openshift-apiserver/apiserver-76f77b778f-w8xn2" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.618598 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9491cb1f-acf6-438b-8175-11737d7bd245-console-serving-cert\") pod \"console-f9d7485db-292nl\" (UID: \"9491cb1f-acf6-438b-8175-11737d7bd245\") " pod="openshift-console/console-f9d7485db-292nl" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.618632 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wggrb\" (UniqueName: \"kubernetes.io/projected/9491cb1f-acf6-438b-8175-11737d7bd245-kube-api-access-wggrb\") pod \"console-f9d7485db-292nl\" (UID: \"9491cb1f-acf6-438b-8175-11737d7bd245\") " pod="openshift-console/console-f9d7485db-292nl" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.618651 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0d0699e2-c1b3-48a3-8d17-5d89424b4bb0-auth-proxy-config\") pod \"machine-approver-56656f9798-wq2zz\" (UID: \"0d0699e2-c1b3-48a3-8d17-5d89424b4bb0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wq2zz" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.618670 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-sg6bg\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.618688 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fttq4\" (UniqueName: \"kubernetes.io/projected/04b9f090-d8f6-4122-bbff-af681fcc4f42-kube-api-access-fttq4\") pod \"csi-hostpathplugin-9nmnq\" (UID: \"04b9f090-d8f6-4122-bbff-af681fcc4f42\") " pod="hostpath-provisioner/csi-hostpathplugin-9nmnq" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.618757 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/757ee90a-18e3-4ccb-b03c-43e90bf8cb34-config\") pod \"etcd-operator-b45778765-xt48p\" (UID: \"757ee90a-18e3-4ccb-b03c-43e90bf8cb34\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xt48p" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.618779 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/952ae5ba-156f-41a7-950a-1783e2e69f94-registry-tls\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.618804 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2404e1f8-203a-4055-8f66-21d8867e7a92-audit-policies\") pod \"oauth-openshift-558db77b4-sg6bg\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.618822 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-sg6bg\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.618843 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/4d793912-71b9-4100-ab1a-e24a0d10dd7e-node-bootstrap-token\") pod \"machine-config-server-vq52n\" (UID: \"4d793912-71b9-4100-ab1a-e24a0d10dd7e\") " pod="openshift-machine-config-operator/machine-config-server-vq52n" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.618866 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/0d0699e2-c1b3-48a3-8d17-5d89424b4bb0-machine-approver-tls\") pod \"machine-approver-56656f9798-wq2zz\" (UID: \"0d0699e2-c1b3-48a3-8d17-5d89424b4bb0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wq2zz" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.618885 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-sg6bg\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.618906 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/04b9f090-d8f6-4122-bbff-af681fcc4f42-socket-dir\") pod \"csi-hostpathplugin-9nmnq\" (UID: \"04b9f090-d8f6-4122-bbff-af681fcc4f42\") " pod="hostpath-provisioner/csi-hostpathplugin-9nmnq" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.618932 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/952ae5ba-156f-41a7-950a-1783e2e69f94-trusted-ca\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.618948 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ae767d9a-3df4-413a-bc25-8fc819c7088e-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-xfb2s\" (UID: \"ae767d9a-3df4-413a-bc25-8fc819c7088e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xfb2s" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.618967 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bq77\" (UniqueName: \"kubernetes.io/projected/23e5ea75-df3a-4a09-87f4-542f0063f4aa-kube-api-access-6bq77\") pod \"ingress-canary-pwc6t\" (UID: \"23e5ea75-df3a-4a09-87f4-542f0063f4aa\") " pod="openshift-ingress-canary/ingress-canary-pwc6t" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.618987 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9491cb1f-acf6-438b-8175-11737d7bd245-oauth-serving-cert\") pod \"console-f9d7485db-292nl\" (UID: \"9491cb1f-acf6-438b-8175-11737d7bd245\") " pod="openshift-console/console-f9d7485db-292nl" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.619004 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/67e07bc1-ce0f-453c-83ba-6f8cc36ed889-metrics-tls\") pod \"dns-operator-744455d44c-f5kts\" (UID: \"67e07bc1-ce0f-453c-83ba-6f8cc36ed889\") " pod="openshift-dns-operator/dns-operator-744455d44c-f5kts" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.619037 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nm59\" (UniqueName: \"kubernetes.io/projected/4d793912-71b9-4100-ab1a-e24a0d10dd7e-kube-api-access-8nm59\") pod \"machine-config-server-vq52n\" (UID: \"4d793912-71b9-4100-ab1a-e24a0d10dd7e\") " pod="openshift-machine-config-operator/machine-config-server-vq52n" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.619055 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3dc1fc6b-9c79-4a78-af19-d49195bc94be-profile-collector-cert\") pod \"catalog-operator-68c6474976-whl8s\" (UID: \"3dc1fc6b-9c79-4a78-af19-d49195bc94be\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-whl8s" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.619079 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgbxj\" (UniqueName: \"kubernetes.io/projected/16511918-7dea-4667-9119-07f0da1b9a84-kube-api-access-wgbxj\") pod \"console-operator-58897d9998-5bhp6\" (UID: \"16511918-7dea-4667-9119-07f0da1b9a84\") " pod="openshift-console-operator/console-operator-58897d9998-5bhp6" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.619096 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d0699e2-c1b3-48a3-8d17-5d89424b4bb0-config\") pod \"machine-approver-56656f9798-wq2zz\" (UID: \"0d0699e2-c1b3-48a3-8d17-5d89424b4bb0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wq2zz" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.619119 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4af31a76-3090-497d-b39f-965c27a27a1c-proxy-tls\") pod \"machine-config-controller-84d6567774-gxhcq\" (UID: \"4af31a76-3090-497d-b39f-965c27a27a1c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gxhcq" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.619152 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/d0c705dd-9608-4b5e-8c91-2511d8427419-image-import-ca\") pod \"apiserver-76f77b778f-w8xn2\" (UID: \"d0c705dd-9608-4b5e-8c91-2511d8427419\") " pod="openshift-apiserver/apiserver-76f77b778f-w8xn2" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.619173 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldzx7\" (UniqueName: \"kubernetes.io/projected/fb6d3b33-6120-40aa-8c4d-01f186d1ab17-kube-api-access-ldzx7\") pod \"kube-storage-version-migrator-operator-b67b599dd-xdpgx\" (UID: \"fb6d3b33-6120-40aa-8c4d-01f186d1ab17\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xdpgx" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.619193 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/952ae5ba-156f-41a7-950a-1783e2e69f94-registry-certificates\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.619216 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5-serving-cert\") pod \"route-controller-manager-6576b87f9c-q4wmp\" (UID: \"83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q4wmp" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.619259 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/90dd16fd-d069-4007-8c82-4811cb343380-tmpfs\") pod \"packageserver-d55dfcdfc-f5rc8\" (UID: \"90dd16fd-d069-4007-8c82-4811cb343380\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f5rc8" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.619277 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f2208bec-45d1-45f0-8700-1218abfa7642-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-ptnr8\" (UID: \"f2208bec-45d1-45f0-8700-1218abfa7642\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ptnr8" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.619282 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f99594fb-adb1-469b-960f-2a56dc3427bd-config\") pod \"kube-apiserver-operator-766d6c64bb-csmdk\" (UID: \"f99594fb-adb1-469b-960f-2a56dc3427bd\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-csmdk" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.619323 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5-client-ca\") pod \"route-controller-manager-6576b87f9c-q4wmp\" (UID: \"83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q4wmp" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.619361 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-sg6bg\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.619394 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0726d810-1988-4eb3-9077-e8892670dc8c-trusted-ca\") pod \"ingress-operator-5b745b69d9-gvkmx\" (UID: \"0726d810-1988-4eb3-9077-e8892670dc8c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-gvkmx" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.619422 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qwvk\" (UniqueName: \"kubernetes.io/projected/83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5-kube-api-access-4qwvk\") pod \"route-controller-manager-6576b87f9c-q4wmp\" (UID: \"83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q4wmp" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.619448 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9491cb1f-acf6-438b-8175-11737d7bd245-trusted-ca-bundle\") pod \"console-f9d7485db-292nl\" (UID: \"9491cb1f-acf6-438b-8175-11737d7bd245\") " pod="openshift-console/console-f9d7485db-292nl" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.619467 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9a6d7d20-0eea-42ae-92e0-91b1b555bf6f-config-volume\") pod \"dns-default-crgj6\" (UID: \"9a6d7d20-0eea-42ae-92e0-91b1b555bf6f\") " pod="openshift-dns/dns-default-crgj6" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.619517 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fd09645-8bbe-458c-8d9b-b4c2875fd963-config\") pod \"kube-controller-manager-operator-78b949d7b-gscbx\" (UID: \"4fd09645-8bbe-458c-8d9b-b4c2875fd963\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gscbx" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.619546 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7aee65fe-0d01-4db7-95ee-131e6ead9071-auth-proxy-config\") pod \"machine-config-operator-74547568cd-s5ct4\" (UID: \"7aee65fe-0d01-4db7-95ee-131e6ead9071\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s5ct4" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.619576 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7aee65fe-0d01-4db7-95ee-131e6ead9071-proxy-tls\") pod \"machine-config-operator-74547568cd-s5ct4\" (UID: \"7aee65fe-0d01-4db7-95ee-131e6ead9071\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s5ct4" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.619607 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d0c705dd-9608-4b5e-8c91-2511d8427419-etcd-serving-ca\") pod \"apiserver-76f77b778f-w8xn2\" (UID: \"d0c705dd-9608-4b5e-8c91-2511d8427419\") " pod="openshift-apiserver/apiserver-76f77b778f-w8xn2" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.619625 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7q6cm\" (UniqueName: \"kubernetes.io/projected/f2208bec-45d1-45f0-8700-1218abfa7642-kube-api-access-7q6cm\") pod \"cluster-image-registry-operator-dc59b4c8b-ptnr8\" (UID: \"f2208bec-45d1-45f0-8700-1218abfa7642\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ptnr8" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.619646 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4fd09645-8bbe-458c-8d9b-b4c2875fd963-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-gscbx\" (UID: \"4fd09645-8bbe-458c-8d9b-b4c2875fd963\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gscbx" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.619705 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/812f722d-c159-40bb-8901-bac36d182280-serving-cert\") pod \"service-ca-operator-777779d784-x9tsn\" (UID: \"812f722d-c159-40bb-8901-bac36d182280\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-x9tsn" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.619725 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpfsm\" (UniqueName: \"kubernetes.io/projected/90dd16fd-d069-4007-8c82-4811cb343380-kube-api-access-zpfsm\") pod \"packageserver-d55dfcdfc-f5rc8\" (UID: \"90dd16fd-d069-4007-8c82-4811cb343380\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f5rc8" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.619754 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d0c705dd-9608-4b5e-8c91-2511d8427419-audit-dir\") pod \"apiserver-76f77b778f-w8xn2\" (UID: \"d0c705dd-9608-4b5e-8c91-2511d8427419\") " pod="openshift-apiserver/apiserver-76f77b778f-w8xn2" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.619779 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f2208bec-45d1-45f0-8700-1218abfa7642-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-ptnr8\" (UID: \"f2208bec-45d1-45f0-8700-1218abfa7642\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ptnr8" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.619800 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-sg6bg\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.619821 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5-config\") pod \"route-controller-manager-6576b87f9c-q4wmp\" (UID: \"83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q4wmp" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.619863 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/757ee90a-18e3-4ccb-b03c-43e90bf8cb34-etcd-service-ca\") pod \"etcd-operator-b45778765-xt48p\" (UID: \"757ee90a-18e3-4ccb-b03c-43e90bf8cb34\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xt48p" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.619886 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9th6b\" (UniqueName: \"kubernetes.io/projected/0d0699e2-c1b3-48a3-8d17-5d89424b4bb0-kube-api-access-9th6b\") pod \"machine-approver-56656f9798-wq2zz\" (UID: \"0d0699e2-c1b3-48a3-8d17-5d89424b4bb0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wq2zz" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.619904 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/04b9f090-d8f6-4122-bbff-af681fcc4f42-registration-dir\") pod \"csi-hostpathplugin-9nmnq\" (UID: \"04b9f090-d8f6-4122-bbff-af681fcc4f42\") " pod="hostpath-provisioner/csi-hostpathplugin-9nmnq" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.619929 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9a6d7d20-0eea-42ae-92e0-91b1b555bf6f-metrics-tls\") pod \"dns-default-crgj6\" (UID: \"9a6d7d20-0eea-42ae-92e0-91b1b555bf6f\") " pod="openshift-dns/dns-default-crgj6" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.619964 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0c705dd-9608-4b5e-8c91-2511d8427419-config\") pod \"apiserver-76f77b778f-w8xn2\" (UID: \"d0c705dd-9608-4b5e-8c91-2511d8427419\") " pod="openshift-apiserver/apiserver-76f77b778f-w8xn2" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.619985 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/90dd16fd-d069-4007-8c82-4811cb343380-webhook-cert\") pod \"packageserver-d55dfcdfc-f5rc8\" (UID: \"90dd16fd-d069-4007-8c82-4811cb343380\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f5rc8" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.620039 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/e686de65-3d86-4826-9f0a-4132c81e9ed2-stats-auth\") pod \"router-default-5444994796-h9kpk\" (UID: \"e686de65-3d86-4826-9f0a-4132c81e9ed2\") " pod="openshift-ingress/router-default-5444994796-h9kpk" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.620071 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6h25\" (UniqueName: \"kubernetes.io/projected/645ef95b-f94d-4ba4-8f10-fbe046a9d9a7-kube-api-access-z6h25\") pod \"openshift-controller-manager-operator-756b6f6bc6-7dcf8\" (UID: \"645ef95b-f94d-4ba4-8f10-fbe046a9d9a7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7dcf8" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.620090 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/17e954cb-2d58-4616-bd5c-e813644d411d-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-qvzhc\" (UID: \"17e954cb-2d58-4616-bd5c-e813644d411d\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-qvzhc" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.620122 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/83688ccb-ace0-4b0a-b584-e650671f5686-config-volume\") pod \"collect-profiles-29409630-lnw2h\" (UID: \"83688ccb-ace0-4b0a-b584-e650671f5686\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-lnw2h" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.620440 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0d0699e2-c1b3-48a3-8d17-5d89424b4bb0-auth-proxy-config\") pod \"machine-approver-56656f9798-wq2zz\" (UID: \"0d0699e2-c1b3-48a3-8d17-5d89424b4bb0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wq2zz" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.620640 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/d0c705dd-9608-4b5e-8c91-2511d8427419-audit\") pod \"apiserver-76f77b778f-w8xn2\" (UID: \"d0c705dd-9608-4b5e-8c91-2511d8427419\") " pod="openshift-apiserver/apiserver-76f77b778f-w8xn2" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.621762 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d0c705dd-9608-4b5e-8c91-2511d8427419-etcd-serving-ca\") pod \"apiserver-76f77b778f-w8xn2\" (UID: \"d0c705dd-9608-4b5e-8c91-2511d8427419\") " pod="openshift-apiserver/apiserver-76f77b778f-w8xn2" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.623325 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-sg6bg\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.623785 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmww7\" (UniqueName: \"kubernetes.io/projected/d0c705dd-9608-4b5e-8c91-2511d8427419-kube-api-access-cmww7\") pod \"apiserver-76f77b778f-w8xn2\" (UID: \"d0c705dd-9608-4b5e-8c91-2511d8427419\") " pod="openshift-apiserver/apiserver-76f77b778f-w8xn2" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.623826 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4af31a76-3090-497d-b39f-965c27a27a1c-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-gxhcq\" (UID: \"4af31a76-3090-497d-b39f-965c27a27a1c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gxhcq" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.623853 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/83688ccb-ace0-4b0a-b584-e650671f5686-secret-volume\") pod \"collect-profiles-29409630-lnw2h\" (UID: \"83688ccb-ace0-4b0a-b584-e650671f5686\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-lnw2h" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.623923 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16511918-7dea-4667-9119-07f0da1b9a84-config\") pod \"console-operator-58897d9998-5bhp6\" (UID: \"16511918-7dea-4667-9119-07f0da1b9a84\") " pod="openshift-console-operator/console-operator-58897d9998-5bhp6" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.624524 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/757ee90a-18e3-4ccb-b03c-43e90bf8cb34-etcd-service-ca\") pod \"etcd-operator-b45778765-xt48p\" (UID: \"757ee90a-18e3-4ccb-b03c-43e90bf8cb34\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xt48p" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.624850 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2404e1f8-203a-4055-8f66-21d8867e7a92-audit-policies\") pod \"oauth-openshift-558db77b4-sg6bg\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.624913 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-sg6bg\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.625649 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/17e954cb-2d58-4616-bd5c-e813644d411d-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-qvzhc\" (UID: \"17e954cb-2d58-4616-bd5c-e813644d411d\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-qvzhc" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.627230 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ae767d9a-3df4-413a-bc25-8fc819c7088e-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-xfb2s\" (UID: \"ae767d9a-3df4-413a-bc25-8fc819c7088e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xfb2s" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.627227 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/952ae5ba-156f-41a7-950a-1783e2e69f94-installation-pull-secrets\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.627288 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/757ee90a-18e3-4ccb-b03c-43e90bf8cb34-etcd-client\") pod \"etcd-operator-b45778765-xt48p\" (UID: \"757ee90a-18e3-4ccb-b03c-43e90bf8cb34\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xt48p" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.627315 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/04b9f090-d8f6-4122-bbff-af681fcc4f42-csi-data-dir\") pod \"csi-hostpathplugin-9nmnq\" (UID: \"04b9f090-d8f6-4122-bbff-af681fcc4f42\") " pod="hostpath-provisioner/csi-hostpathplugin-9nmnq" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.627342 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/16511918-7dea-4667-9119-07f0da1b9a84-serving-cert\") pod \"console-operator-58897d9998-5bhp6\" (UID: \"16511918-7dea-4667-9119-07f0da1b9a84\") " pod="openshift-console-operator/console-operator-58897d9998-5bhp6" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.627364 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/757ee90a-18e3-4ccb-b03c-43e90bf8cb34-etcd-ca\") pod \"etcd-operator-b45778765-xt48p\" (UID: \"757ee90a-18e3-4ccb-b03c-43e90bf8cb34\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xt48p" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.627387 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dd6b\" (UniqueName: \"kubernetes.io/projected/2a86c401-c7dd-4b95-b267-8c58e5728e6a-kube-api-access-8dd6b\") pod \"package-server-manager-789f6589d5-mvczl\" (UID: \"2a86c401-c7dd-4b95-b267-8c58e5728e6a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mvczl" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.628322 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/757ee90a-18e3-4ccb-b03c-43e90bf8cb34-etcd-ca\") pod \"etcd-operator-b45778765-xt48p\" (UID: \"757ee90a-18e3-4ccb-b03c-43e90bf8cb34\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xt48p" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.628536 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/d0c705dd-9608-4b5e-8c91-2511d8427419-image-import-ca\") pod \"apiserver-76f77b778f-w8xn2\" (UID: \"d0c705dd-9608-4b5e-8c91-2511d8427419\") " pod="openshift-apiserver/apiserver-76f77b778f-w8xn2" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.629088 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d0c705dd-9608-4b5e-8c91-2511d8427419-etcd-client\") pod \"apiserver-76f77b778f-w8xn2\" (UID: \"d0c705dd-9608-4b5e-8c91-2511d8427419\") " pod="openshift-apiserver/apiserver-76f77b778f-w8xn2" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.629217 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-sg6bg\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.629282 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-sg6bg\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.629329 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/23e5ea75-df3a-4a09-87f4-542f0063f4aa-cert\") pod \"ingress-canary-pwc6t\" (UID: \"23e5ea75-df3a-4a09-87f4-542f0063f4aa\") " pod="openshift-ingress-canary/ingress-canary-pwc6t" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.629365 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72gzp\" (UniqueName: \"kubernetes.io/projected/83688ccb-ace0-4b0a-b584-e650671f5686-kube-api-access-72gzp\") pod \"collect-profiles-29409630-lnw2h\" (UID: \"83688ccb-ace0-4b0a-b584-e650671f5686\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-lnw2h" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.629468 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/e686de65-3d86-4826-9f0a-4132c81e9ed2-default-certificate\") pod \"router-default-5444994796-h9kpk\" (UID: \"e686de65-3d86-4826-9f0a-4132c81e9ed2\") " pod="openshift-ingress/router-default-5444994796-h9kpk" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.629509 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/16511918-7dea-4667-9119-07f0da1b9a84-trusted-ca\") pod \"console-operator-58897d9998-5bhp6\" (UID: \"16511918-7dea-4667-9119-07f0da1b9a84\") " pod="openshift-console-operator/console-operator-58897d9998-5bhp6" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.629586 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25wjr\" (UniqueName: \"kubernetes.io/projected/4af31a76-3090-497d-b39f-965c27a27a1c-kube-api-access-25wjr\") pod \"machine-config-controller-84d6567774-gxhcq\" (UID: \"4af31a76-3090-497d-b39f-965c27a27a1c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gxhcq" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.629624 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwvvt\" (UniqueName: \"kubernetes.io/projected/10d7dabb-aa6f-4f2b-acf9-b02dc369d8be-kube-api-access-rwvvt\") pod \"cluster-samples-operator-665b6dd947-nd5kp\" (UID: \"10d7dabb-aa6f-4f2b-acf9-b02dc369d8be\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-nd5kp" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.629655 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-sg6bg\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.629683 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72t8c\" (UniqueName: \"kubernetes.io/projected/2ea695f7-4e9e-4b12-8e0c-22431291576d-kube-api-access-72t8c\") pod \"marketplace-operator-79b997595-7hstm\" (UID: \"2ea695f7-4e9e-4b12-8e0c-22431291576d\") " pod="openshift-marketplace/marketplace-operator-79b997595-7hstm" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.629716 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d0c705dd-9608-4b5e-8c91-2511d8427419-serving-cert\") pod \"apiserver-76f77b778f-w8xn2\" (UID: \"d0c705dd-9608-4b5e-8c91-2511d8427419\") " pod="openshift-apiserver/apiserver-76f77b778f-w8xn2" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.629741 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0726d810-1988-4eb3-9077-e8892670dc8c-bound-sa-token\") pod \"ingress-operator-5b745b69d9-gvkmx\" (UID: \"0726d810-1988-4eb3-9077-e8892670dc8c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-gvkmx" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.629773 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9491cb1f-acf6-438b-8175-11737d7bd245-console-oauth-config\") pod \"console-f9d7485db-292nl\" (UID: \"9491cb1f-acf6-438b-8175-11737d7bd245\") " pod="openshift-console/console-f9d7485db-292nl" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.629889 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/90dd16fd-d069-4007-8c82-4811cb343380-apiservice-cert\") pod \"packageserver-d55dfcdfc-f5rc8\" (UID: \"90dd16fd-d069-4007-8c82-4811cb343380\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f5rc8" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.629927 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnvrw\" (UniqueName: \"kubernetes.io/projected/7aee65fe-0d01-4db7-95ee-131e6ead9071-kube-api-access-lnvrw\") pod \"machine-config-operator-74547568cd-s5ct4\" (UID: \"7aee65fe-0d01-4db7-95ee-131e6ead9071\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s5ct4" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.629958 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xv9q\" (UniqueName: \"kubernetes.io/projected/2404e1f8-203a-4055-8f66-21d8867e7a92-kube-api-access-4xv9q\") pod \"oauth-openshift-558db77b4-sg6bg\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.629983 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/04b9f090-d8f6-4122-bbff-af681fcc4f42-plugins-dir\") pod \"csi-hostpathplugin-9nmnq\" (UID: \"04b9f090-d8f6-4122-bbff-af681fcc4f42\") " pod="hostpath-provisioner/csi-hostpathplugin-9nmnq" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.630007 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e686de65-3d86-4826-9f0a-4132c81e9ed2-metrics-certs\") pod \"router-default-5444994796-h9kpk\" (UID: \"e686de65-3d86-4826-9f0a-4132c81e9ed2\") " pod="openshift-ingress/router-default-5444994796-h9kpk" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.630046 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cbnx\" (UniqueName: \"kubernetes.io/projected/e686de65-3d86-4826-9f0a-4132c81e9ed2-kube-api-access-2cbnx\") pod \"router-default-5444994796-h9kpk\" (UID: \"e686de65-3d86-4826-9f0a-4132c81e9ed2\") " pod="openshift-ingress/router-default-5444994796-h9kpk" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.630073 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2ea695f7-4e9e-4b12-8e0c-22431291576d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7hstm\" (UID: \"2ea695f7-4e9e-4b12-8e0c-22431291576d\") " pod="openshift-marketplace/marketplace-operator-79b997595-7hstm" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.630111 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.630137 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/952ae5ba-156f-41a7-950a-1783e2e69f94-bound-sa-token\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.630170 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z78ht\" (UniqueName: \"kubernetes.io/projected/17e954cb-2d58-4616-bd5c-e813644d411d-kube-api-access-z78ht\") pod \"multus-admission-controller-857f4d67dd-qvzhc\" (UID: \"17e954cb-2d58-4616-bd5c-e813644d411d\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-qvzhc" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.630194 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d0c705dd-9608-4b5e-8c91-2511d8427419-encryption-config\") pod \"apiserver-76f77b778f-w8xn2\" (UID: \"d0c705dd-9608-4b5e-8c91-2511d8427419\") " pod="openshift-apiserver/apiserver-76f77b778f-w8xn2" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.630551 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9491cb1f-acf6-438b-8175-11737d7bd245-trusted-ca-bundle\") pod \"console-f9d7485db-292nl\" (UID: \"9491cb1f-acf6-438b-8175-11737d7bd245\") " pod="openshift-console/console-f9d7485db-292nl" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.632497 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-sg6bg\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.632619 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/757ee90a-18e3-4ccb-b03c-43e90bf8cb34-etcd-client\") pod \"etcd-operator-b45778765-xt48p\" (UID: \"757ee90a-18e3-4ccb-b03c-43e90bf8cb34\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xt48p" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.632626 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5-serving-cert\") pod \"route-controller-manager-6576b87f9c-q4wmp\" (UID: \"83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q4wmp" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.630601 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d0c705dd-9608-4b5e-8c91-2511d8427419-audit-dir\") pod \"apiserver-76f77b778f-w8xn2\" (UID: \"d0c705dd-9608-4b5e-8c91-2511d8427419\") " pod="openshift-apiserver/apiserver-76f77b778f-w8xn2" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.632992 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-sg6bg\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.633137 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/4d793912-71b9-4100-ab1a-e24a0d10dd7e-certs\") pod \"machine-config-server-vq52n\" (UID: \"4d793912-71b9-4100-ab1a-e24a0d10dd7e\") " pod="openshift-machine-config-operator/machine-config-server-vq52n" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.633178 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5-config\") pod \"route-controller-manager-6576b87f9c-q4wmp\" (UID: \"83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q4wmp" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.633187 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/952ae5ba-156f-41a7-950a-1783e2e69f94-ca-trust-extracted\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.633271 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/10d7dabb-aa6f-4f2b-acf9-b02dc369d8be-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-nd5kp\" (UID: \"10d7dabb-aa6f-4f2b-acf9-b02dc369d8be\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-nd5kp" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.633306 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d0c705dd-9608-4b5e-8c91-2511d8427419-node-pullsecrets\") pod \"apiserver-76f77b778f-w8xn2\" (UID: \"d0c705dd-9608-4b5e-8c91-2511d8427419\") " pod="openshift-apiserver/apiserver-76f77b778f-w8xn2" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.633352 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/645ef95b-f94d-4ba4-8f10-fbe046a9d9a7-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-7dcf8\" (UID: \"645ef95b-f94d-4ba4-8f10-fbe046a9d9a7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7dcf8" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.633428 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-sg6bg\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.633503 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/12ddf317-c034-4154-9966-c8b502ec0d6c-signing-key\") pod \"service-ca-9c57cc56f-pp9bw\" (UID: \"12ddf317-c034-4154-9966-c8b502ec0d6c\") " pod="openshift-service-ca/service-ca-9c57cc56f-pp9bw" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.633568 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/12ddf317-c034-4154-9966-c8b502ec0d6c-signing-cabundle\") pod \"service-ca-9c57cc56f-pp9bw\" (UID: \"12ddf317-c034-4154-9966-c8b502ec0d6c\") " pod="openshift-service-ca/service-ca-9c57cc56f-pp9bw" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.633649 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d0c705dd-9608-4b5e-8c91-2511d8427419-trusted-ca-bundle\") pod \"apiserver-76f77b778f-w8xn2\" (UID: \"d0c705dd-9608-4b5e-8c91-2511d8427419\") " pod="openshift-apiserver/apiserver-76f77b778f-w8xn2" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.633646 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/952ae5ba-156f-41a7-950a-1783e2e69f94-ca-trust-extracted\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.633717 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/952ae5ba-156f-41a7-950a-1783e2e69f94-trusted-ca\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.633731 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4fd09645-8bbe-458c-8d9b-b4c2875fd963-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-gscbx\" (UID: \"4fd09645-8bbe-458c-8d9b-b4c2875fd963\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gscbx" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.633006 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/0d0699e2-c1b3-48a3-8d17-5d89424b4bb0-machine-approver-tls\") pod \"machine-approver-56656f9798-wq2zz\" (UID: \"0d0699e2-c1b3-48a3-8d17-5d89424b4bb0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wq2zz" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.634246 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/757ee90a-18e3-4ccb-b03c-43e90bf8cb34-serving-cert\") pod \"etcd-operator-b45778765-xt48p\" (UID: \"757ee90a-18e3-4ccb-b03c-43e90bf8cb34\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xt48p" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.634455 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/952ae5ba-156f-41a7-950a-1783e2e69f94-registry-certificates\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.634608 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9491cb1f-acf6-438b-8175-11737d7bd245-console-serving-cert\") pod \"console-f9d7485db-292nl\" (UID: \"9491cb1f-acf6-438b-8175-11737d7bd245\") " pod="openshift-console/console-f9d7485db-292nl" Dec 01 08:42:40 crc kubenswrapper[4873]: E1201 08:42:40.635750 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:41.135530924 +0000 UTC m=+137.037639463 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.638929 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d0c705dd-9608-4b5e-8c91-2511d8427419-node-pullsecrets\") pod \"apiserver-76f77b778f-w8xn2\" (UID: \"d0c705dd-9608-4b5e-8c91-2511d8427419\") " pod="openshift-apiserver/apiserver-76f77b778f-w8xn2" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.639596 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/952ae5ba-156f-41a7-950a-1783e2e69f94-registry-tls\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.639700 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0726d810-1988-4eb3-9077-e8892670dc8c-metrics-tls\") pod \"ingress-operator-5b745b69d9-gvkmx\" (UID: \"0726d810-1988-4eb3-9077-e8892670dc8c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-gvkmx" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.639786 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlc6g\" (UniqueName: \"kubernetes.io/projected/952ae5ba-156f-41a7-950a-1783e2e69f94-kube-api-access-qlc6g\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.639822 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwdlq\" (UniqueName: \"kubernetes.io/projected/757ee90a-18e3-4ccb-b03c-43e90bf8cb34-kube-api-access-gwdlq\") pod \"etcd-operator-b45778765-xt48p\" (UID: \"757ee90a-18e3-4ccb-b03c-43e90bf8cb34\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xt48p" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.639919 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2404e1f8-203a-4055-8f66-21d8867e7a92-audit-dir\") pod \"oauth-openshift-558db77b4-sg6bg\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.639979 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f99594fb-adb1-469b-960f-2a56dc3427bd-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-csmdk\" (UID: \"f99594fb-adb1-469b-960f-2a56dc3427bd\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-csmdk" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.640145 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-sg6bg\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.640399 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/645ef95b-f94d-4ba4-8f10-fbe046a9d9a7-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-7dcf8\" (UID: \"645ef95b-f94d-4ba4-8f10-fbe046a9d9a7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7dcf8" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.640447 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9491cb1f-acf6-438b-8175-11737d7bd245-service-ca\") pod \"console-f9d7485db-292nl\" (UID: \"9491cb1f-acf6-438b-8175-11737d7bd245\") " pod="openshift-console/console-f9d7485db-292nl" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.640489 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dcn4\" (UniqueName: \"kubernetes.io/projected/67e07bc1-ce0f-453c-83ba-6f8cc36ed889-kube-api-access-9dcn4\") pod \"dns-operator-744455d44c-f5kts\" (UID: \"67e07bc1-ce0f-453c-83ba-6f8cc36ed889\") " pod="openshift-dns-operator/dns-operator-744455d44c-f5kts" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.640635 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c92cq\" (UniqueName: \"kubernetes.io/projected/ae767d9a-3df4-413a-bc25-8fc819c7088e-kube-api-access-c92cq\") pod \"openshift-apiserver-operator-796bbdcf4f-xfb2s\" (UID: \"ae767d9a-3df4-413a-bc25-8fc819c7088e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xfb2s" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.640812 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e686de65-3d86-4826-9f0a-4132c81e9ed2-service-ca-bundle\") pod \"router-default-5444994796-h9kpk\" (UID: \"e686de65-3d86-4826-9f0a-4132c81e9ed2\") " pod="openshift-ingress/router-default-5444994796-h9kpk" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.641916 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-sg6bg\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.642159 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d0c705dd-9608-4b5e-8c91-2511d8427419-serving-cert\") pod \"apiserver-76f77b778f-w8xn2\" (UID: \"d0c705dd-9608-4b5e-8c91-2511d8427419\") " pod="openshift-apiserver/apiserver-76f77b778f-w8xn2" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.642399 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f2208bec-45d1-45f0-8700-1218abfa7642-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-ptnr8\" (UID: \"f2208bec-45d1-45f0-8700-1218abfa7642\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ptnr8" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.642723 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/952ae5ba-156f-41a7-950a-1783e2e69f94-installation-pull-secrets\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.643923 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9491cb1f-acf6-438b-8175-11737d7bd245-service-ca\") pod \"console-f9d7485db-292nl\" (UID: \"9491cb1f-acf6-438b-8175-11737d7bd245\") " pod="openshift-console/console-f9d7485db-292nl" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.643948 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d0c705dd-9608-4b5e-8c91-2511d8427419-encryption-config\") pod \"apiserver-76f77b778f-w8xn2\" (UID: \"d0c705dd-9608-4b5e-8c91-2511d8427419\") " pod="openshift-apiserver/apiserver-76f77b778f-w8xn2" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.644443 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2404e1f8-203a-4055-8f66-21d8867e7a92-audit-dir\") pod \"oauth-openshift-558db77b4-sg6bg\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.646153 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-sg6bg\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.650228 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-sg6bg\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.650541 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-sg6bg\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.652974 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-sg6bg\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.654990 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d0699e2-c1b3-48a3-8d17-5d89424b4bb0-config\") pod \"machine-approver-56656f9798-wq2zz\" (UID: \"0d0699e2-c1b3-48a3-8d17-5d89424b4bb0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wq2zz" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.655543 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9491cb1f-acf6-438b-8175-11737d7bd245-console-oauth-config\") pod \"console-f9d7485db-292nl\" (UID: \"9491cb1f-acf6-438b-8175-11737d7bd245\") " pod="openshift-console/console-f9d7485db-292nl" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.655849 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9491cb1f-acf6-438b-8175-11737d7bd245-oauth-serving-cert\") pod \"console-f9d7485db-292nl\" (UID: \"9491cb1f-acf6-438b-8175-11737d7bd245\") " pod="openshift-console/console-f9d7485db-292nl" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.655904 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-sg6bg\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.655964 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fd09645-8bbe-458c-8d9b-b4c2875fd963-config\") pod \"kube-controller-manager-operator-78b949d7b-gscbx\" (UID: \"4fd09645-8bbe-458c-8d9b-b4c2875fd963\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gscbx" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.656541 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/645ef95b-f94d-4ba4-8f10-fbe046a9d9a7-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-7dcf8\" (UID: \"645ef95b-f94d-4ba4-8f10-fbe046a9d9a7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7dcf8" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.656833 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4af31a76-3090-497d-b39f-965c27a27a1c-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-gxhcq\" (UID: \"4af31a76-3090-497d-b39f-965c27a27a1c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gxhcq" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.658874 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-sg6bg\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.662046 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/16511918-7dea-4667-9119-07f0da1b9a84-serving-cert\") pod \"console-operator-58897d9998-5bhp6\" (UID: \"16511918-7dea-4667-9119-07f0da1b9a84\") " pod="openshift-console-operator/console-operator-58897d9998-5bhp6" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.662355 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0c705dd-9608-4b5e-8c91-2511d8427419-config\") pod \"apiserver-76f77b778f-w8xn2\" (UID: \"d0c705dd-9608-4b5e-8c91-2511d8427419\") " pod="openshift-apiserver/apiserver-76f77b778f-w8xn2" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.662452 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/16511918-7dea-4667-9119-07f0da1b9a84-trusted-ca\") pod \"console-operator-58897d9998-5bhp6\" (UID: \"16511918-7dea-4667-9119-07f0da1b9a84\") " pod="openshift-console-operator/console-operator-58897d9998-5bhp6" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.662459 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-sg6bg\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.662581 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d0c705dd-9608-4b5e-8c91-2511d8427419-trusted-ca-bundle\") pod \"apiserver-76f77b778f-w8xn2\" (UID: \"d0c705dd-9608-4b5e-8c91-2511d8427419\") " pod="openshift-apiserver/apiserver-76f77b778f-w8xn2" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.662654 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16511918-7dea-4667-9119-07f0da1b9a84-config\") pod \"console-operator-58897d9998-5bhp6\" (UID: \"16511918-7dea-4667-9119-07f0da1b9a84\") " pod="openshift-console-operator/console-operator-58897d9998-5bhp6" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.665399 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4af31a76-3090-497d-b39f-965c27a27a1c-proxy-tls\") pod \"machine-config-controller-84d6567774-gxhcq\" (UID: \"4af31a76-3090-497d-b39f-965c27a27a1c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gxhcq" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.666077 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f2208bec-45d1-45f0-8700-1218abfa7642-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-ptnr8\" (UID: \"f2208bec-45d1-45f0-8700-1218abfa7642\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ptnr8" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.671473 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/645ef95b-f94d-4ba4-8f10-fbe046a9d9a7-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-7dcf8\" (UID: \"645ef95b-f94d-4ba4-8f10-fbe046a9d9a7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7dcf8" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.686716 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wggrb\" (UniqueName: \"kubernetes.io/projected/9491cb1f-acf6-438b-8175-11737d7bd245-kube-api-access-wggrb\") pod \"console-f9d7485db-292nl\" (UID: \"9491cb1f-acf6-438b-8175-11737d7bd245\") " pod="openshift-console/console-f9d7485db-292nl" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.708509 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4fd09645-8bbe-458c-8d9b-b4c2875fd963-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-gscbx\" (UID: \"4fd09645-8bbe-458c-8d9b-b4c2875fd963\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gscbx" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.730205 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6h25\" (UniqueName: \"kubernetes.io/projected/645ef95b-f94d-4ba4-8f10-fbe046a9d9a7-kube-api-access-z6h25\") pod \"openshift-controller-manager-operator-756b6f6bc6-7dcf8\" (UID: \"645ef95b-f94d-4ba4-8f10-fbe046a9d9a7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7dcf8" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.741847 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.742136 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dcn4\" (UniqueName: \"kubernetes.io/projected/67e07bc1-ce0f-453c-83ba-6f8cc36ed889-kube-api-access-9dcn4\") pod \"dns-operator-744455d44c-f5kts\" (UID: \"67e07bc1-ce0f-453c-83ba-6f8cc36ed889\") " pod="openshift-dns-operator/dns-operator-744455d44c-f5kts" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.742162 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e686de65-3d86-4826-9f0a-4132c81e9ed2-service-ca-bundle\") pod \"router-default-5444994796-h9kpk\" (UID: \"e686de65-3d86-4826-9f0a-4132c81e9ed2\") " pod="openshift-ingress/router-default-5444994796-h9kpk" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.742196 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/812f722d-c159-40bb-8901-bac36d182280-config\") pod \"service-ca-operator-777779d784-x9tsn\" (UID: \"812f722d-c159-40bb-8901-bac36d182280\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-x9tsn" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.742213 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8xhk\" (UniqueName: \"kubernetes.io/projected/e3e97e3b-f082-4093-b183-fcd5ac857c07-kube-api-access-n8xhk\") pod \"olm-operator-6b444d44fb-xrgx6\" (UID: \"e3e97e3b-f082-4093-b183-fcd5ac857c07\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xrgx6" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.742230 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e3e97e3b-f082-4093-b183-fcd5ac857c07-srv-cert\") pod \"olm-operator-6b444d44fb-xrgx6\" (UID: \"e3e97e3b-f082-4093-b183-fcd5ac857c07\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xrgx6" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.742247 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f99594fb-adb1-469b-960f-2a56dc3427bd-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-csmdk\" (UID: \"f99594fb-adb1-469b-960f-2a56dc3427bd\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-csmdk" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.742268 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25z7h\" (UniqueName: \"kubernetes.io/projected/dfbcf0e0-e5a1-41e0-813c-008f355a1f0a-kube-api-access-25z7h\") pod \"migrator-59844c95c7-l9mq8\" (UID: \"dfbcf0e0-e5a1-41e0-813c-008f355a1f0a\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-l9mq8" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.742284 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb6d3b33-6120-40aa-8c4d-01f186d1ab17-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-xdpgx\" (UID: \"fb6d3b33-6120-40aa-8c4d-01f186d1ab17\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xdpgx" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.742299 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e3e97e3b-f082-4093-b183-fcd5ac857c07-profile-collector-cert\") pod \"olm-operator-6b444d44fb-xrgx6\" (UID: \"e3e97e3b-f082-4093-b183-fcd5ac857c07\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xrgx6" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.742315 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2ea695f7-4e9e-4b12-8e0c-22431291576d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7hstm\" (UID: \"2ea695f7-4e9e-4b12-8e0c-22431291576d\") " pod="openshift-marketplace/marketplace-operator-79b997595-7hstm" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.742332 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrqmg\" (UniqueName: \"kubernetes.io/projected/12ddf317-c034-4154-9966-c8b502ec0d6c-kube-api-access-nrqmg\") pod \"service-ca-9c57cc56f-pp9bw\" (UID: \"12ddf317-c034-4154-9966-c8b502ec0d6c\") " pod="openshift-service-ca/service-ca-9c57cc56f-pp9bw" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.742353 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7aee65fe-0d01-4db7-95ee-131e6ead9071-images\") pod \"machine-config-operator-74547568cd-s5ct4\" (UID: \"7aee65fe-0d01-4db7-95ee-131e6ead9071\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s5ct4" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.742398 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fb6d3b33-6120-40aa-8c4d-01f186d1ab17-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-xdpgx\" (UID: \"fb6d3b33-6120-40aa-8c4d-01f186d1ab17\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xdpgx" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.742416 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cp6l\" (UniqueName: \"kubernetes.io/projected/9a6d7d20-0eea-42ae-92e0-91b1b555bf6f-kube-api-access-8cp6l\") pod \"dns-default-crgj6\" (UID: \"9a6d7d20-0eea-42ae-92e0-91b1b555bf6f\") " pod="openshift-dns/dns-default-crgj6" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.742432 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3dc1fc6b-9c79-4a78-af19-d49195bc94be-srv-cert\") pod \"catalog-operator-68c6474976-whl8s\" (UID: \"3dc1fc6b-9c79-4a78-af19-d49195bc94be\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-whl8s" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.742449 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/04b9f090-d8f6-4122-bbff-af681fcc4f42-mountpoint-dir\") pod \"csi-hostpathplugin-9nmnq\" (UID: \"04b9f090-d8f6-4122-bbff-af681fcc4f42\") " pod="hostpath-provisioner/csi-hostpathplugin-9nmnq" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.742468 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/2a86c401-c7dd-4b95-b267-8c58e5728e6a-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-mvczl\" (UID: \"2a86c401-c7dd-4b95-b267-8c58e5728e6a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mvczl" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.742486 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkwmv\" (UniqueName: \"kubernetes.io/projected/3dc1fc6b-9c79-4a78-af19-d49195bc94be-kube-api-access-mkwmv\") pod \"catalog-operator-68c6474976-whl8s\" (UID: \"3dc1fc6b-9c79-4a78-af19-d49195bc94be\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-whl8s" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.742503 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7n4g\" (UniqueName: \"kubernetes.io/projected/0726d810-1988-4eb3-9077-e8892670dc8c-kube-api-access-z7n4g\") pod \"ingress-operator-5b745b69d9-gvkmx\" (UID: \"0726d810-1988-4eb3-9077-e8892670dc8c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-gvkmx" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.742520 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7hln\" (UniqueName: \"kubernetes.io/projected/812f722d-c159-40bb-8901-bac36d182280-kube-api-access-w7hln\") pod \"service-ca-operator-777779d784-x9tsn\" (UID: \"812f722d-c159-40bb-8901-bac36d182280\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-x9tsn" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.742550 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fttq4\" (UniqueName: \"kubernetes.io/projected/04b9f090-d8f6-4122-bbff-af681fcc4f42-kube-api-access-fttq4\") pod \"csi-hostpathplugin-9nmnq\" (UID: \"04b9f090-d8f6-4122-bbff-af681fcc4f42\") " pod="hostpath-provisioner/csi-hostpathplugin-9nmnq" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.742566 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/4d793912-71b9-4100-ab1a-e24a0d10dd7e-node-bootstrap-token\") pod \"machine-config-server-vq52n\" (UID: \"4d793912-71b9-4100-ab1a-e24a0d10dd7e\") " pod="openshift-machine-config-operator/machine-config-server-vq52n" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.742584 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bq77\" (UniqueName: \"kubernetes.io/projected/23e5ea75-df3a-4a09-87f4-542f0063f4aa-kube-api-access-6bq77\") pod \"ingress-canary-pwc6t\" (UID: \"23e5ea75-df3a-4a09-87f4-542f0063f4aa\") " pod="openshift-ingress-canary/ingress-canary-pwc6t" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.742600 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/04b9f090-d8f6-4122-bbff-af681fcc4f42-socket-dir\") pod \"csi-hostpathplugin-9nmnq\" (UID: \"04b9f090-d8f6-4122-bbff-af681fcc4f42\") " pod="hostpath-provisioner/csi-hostpathplugin-9nmnq" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.742625 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/67e07bc1-ce0f-453c-83ba-6f8cc36ed889-metrics-tls\") pod \"dns-operator-744455d44c-f5kts\" (UID: \"67e07bc1-ce0f-453c-83ba-6f8cc36ed889\") " pod="openshift-dns-operator/dns-operator-744455d44c-f5kts" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.742647 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nm59\" (UniqueName: \"kubernetes.io/projected/4d793912-71b9-4100-ab1a-e24a0d10dd7e-kube-api-access-8nm59\") pod \"machine-config-server-vq52n\" (UID: \"4d793912-71b9-4100-ab1a-e24a0d10dd7e\") " pod="openshift-machine-config-operator/machine-config-server-vq52n" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.742662 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3dc1fc6b-9c79-4a78-af19-d49195bc94be-profile-collector-cert\") pod \"catalog-operator-68c6474976-whl8s\" (UID: \"3dc1fc6b-9c79-4a78-af19-d49195bc94be\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-whl8s" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.742697 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldzx7\" (UniqueName: \"kubernetes.io/projected/fb6d3b33-6120-40aa-8c4d-01f186d1ab17-kube-api-access-ldzx7\") pod \"kube-storage-version-migrator-operator-b67b599dd-xdpgx\" (UID: \"fb6d3b33-6120-40aa-8c4d-01f186d1ab17\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xdpgx" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.742721 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/90dd16fd-d069-4007-8c82-4811cb343380-tmpfs\") pod \"packageserver-d55dfcdfc-f5rc8\" (UID: \"90dd16fd-d069-4007-8c82-4811cb343380\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f5rc8" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.742739 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f99594fb-adb1-469b-960f-2a56dc3427bd-config\") pod \"kube-apiserver-operator-766d6c64bb-csmdk\" (UID: \"f99594fb-adb1-469b-960f-2a56dc3427bd\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-csmdk" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.742758 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0726d810-1988-4eb3-9077-e8892670dc8c-trusted-ca\") pod \"ingress-operator-5b745b69d9-gvkmx\" (UID: \"0726d810-1988-4eb3-9077-e8892670dc8c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-gvkmx" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.742781 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9a6d7d20-0eea-42ae-92e0-91b1b555bf6f-config-volume\") pod \"dns-default-crgj6\" (UID: \"9a6d7d20-0eea-42ae-92e0-91b1b555bf6f\") " pod="openshift-dns/dns-default-crgj6" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.742801 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7aee65fe-0d01-4db7-95ee-131e6ead9071-auth-proxy-config\") pod \"machine-config-operator-74547568cd-s5ct4\" (UID: \"7aee65fe-0d01-4db7-95ee-131e6ead9071\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s5ct4" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.742823 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7aee65fe-0d01-4db7-95ee-131e6ead9071-proxy-tls\") pod \"machine-config-operator-74547568cd-s5ct4\" (UID: \"7aee65fe-0d01-4db7-95ee-131e6ead9071\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s5ct4" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.742843 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpfsm\" (UniqueName: \"kubernetes.io/projected/90dd16fd-d069-4007-8c82-4811cb343380-kube-api-access-zpfsm\") pod \"packageserver-d55dfcdfc-f5rc8\" (UID: \"90dd16fd-d069-4007-8c82-4811cb343380\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f5rc8" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.742860 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/812f722d-c159-40bb-8901-bac36d182280-serving-cert\") pod \"service-ca-operator-777779d784-x9tsn\" (UID: \"812f722d-c159-40bb-8901-bac36d182280\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-x9tsn" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.742876 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/04b9f090-d8f6-4122-bbff-af681fcc4f42-registration-dir\") pod \"csi-hostpathplugin-9nmnq\" (UID: \"04b9f090-d8f6-4122-bbff-af681fcc4f42\") " pod="hostpath-provisioner/csi-hostpathplugin-9nmnq" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.742907 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/90dd16fd-d069-4007-8c82-4811cb343380-webhook-cert\") pod \"packageserver-d55dfcdfc-f5rc8\" (UID: \"90dd16fd-d069-4007-8c82-4811cb343380\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f5rc8" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.742925 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9a6d7d20-0eea-42ae-92e0-91b1b555bf6f-metrics-tls\") pod \"dns-default-crgj6\" (UID: \"9a6d7d20-0eea-42ae-92e0-91b1b555bf6f\") " pod="openshift-dns/dns-default-crgj6" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.742942 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/e686de65-3d86-4826-9f0a-4132c81e9ed2-stats-auth\") pod \"router-default-5444994796-h9kpk\" (UID: \"e686de65-3d86-4826-9f0a-4132c81e9ed2\") " pod="openshift-ingress/router-default-5444994796-h9kpk" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.742964 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/83688ccb-ace0-4b0a-b584-e650671f5686-config-volume\") pod \"collect-profiles-29409630-lnw2h\" (UID: \"83688ccb-ace0-4b0a-b584-e650671f5686\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-lnw2h" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.742989 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/83688ccb-ace0-4b0a-b584-e650671f5686-secret-volume\") pod \"collect-profiles-29409630-lnw2h\" (UID: \"83688ccb-ace0-4b0a-b584-e650671f5686\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-lnw2h" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.743034 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/04b9f090-d8f6-4122-bbff-af681fcc4f42-csi-data-dir\") pod \"csi-hostpathplugin-9nmnq\" (UID: \"04b9f090-d8f6-4122-bbff-af681fcc4f42\") " pod="hostpath-provisioner/csi-hostpathplugin-9nmnq" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.743056 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dd6b\" (UniqueName: \"kubernetes.io/projected/2a86c401-c7dd-4b95-b267-8c58e5728e6a-kube-api-access-8dd6b\") pod \"package-server-manager-789f6589d5-mvczl\" (UID: \"2a86c401-c7dd-4b95-b267-8c58e5728e6a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mvczl" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.743076 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/23e5ea75-df3a-4a09-87f4-542f0063f4aa-cert\") pod \"ingress-canary-pwc6t\" (UID: \"23e5ea75-df3a-4a09-87f4-542f0063f4aa\") " pod="openshift-ingress-canary/ingress-canary-pwc6t" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.743096 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72gzp\" (UniqueName: \"kubernetes.io/projected/83688ccb-ace0-4b0a-b584-e650671f5686-kube-api-access-72gzp\") pod \"collect-profiles-29409630-lnw2h\" (UID: \"83688ccb-ace0-4b0a-b584-e650671f5686\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-lnw2h" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.743115 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/e686de65-3d86-4826-9f0a-4132c81e9ed2-default-certificate\") pod \"router-default-5444994796-h9kpk\" (UID: \"e686de65-3d86-4826-9f0a-4132c81e9ed2\") " pod="openshift-ingress/router-default-5444994796-h9kpk" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.743147 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72t8c\" (UniqueName: \"kubernetes.io/projected/2ea695f7-4e9e-4b12-8e0c-22431291576d-kube-api-access-72t8c\") pod \"marketplace-operator-79b997595-7hstm\" (UID: \"2ea695f7-4e9e-4b12-8e0c-22431291576d\") " pod="openshift-marketplace/marketplace-operator-79b997595-7hstm" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.743164 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0726d810-1988-4eb3-9077-e8892670dc8c-bound-sa-token\") pod \"ingress-operator-5b745b69d9-gvkmx\" (UID: \"0726d810-1988-4eb3-9077-e8892670dc8c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-gvkmx" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.743182 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/90dd16fd-d069-4007-8c82-4811cb343380-apiservice-cert\") pod \"packageserver-d55dfcdfc-f5rc8\" (UID: \"90dd16fd-d069-4007-8c82-4811cb343380\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f5rc8" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.743200 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnvrw\" (UniqueName: \"kubernetes.io/projected/7aee65fe-0d01-4db7-95ee-131e6ead9071-kube-api-access-lnvrw\") pod \"machine-config-operator-74547568cd-s5ct4\" (UID: \"7aee65fe-0d01-4db7-95ee-131e6ead9071\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s5ct4" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.743219 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/04b9f090-d8f6-4122-bbff-af681fcc4f42-plugins-dir\") pod \"csi-hostpathplugin-9nmnq\" (UID: \"04b9f090-d8f6-4122-bbff-af681fcc4f42\") " pod="hostpath-provisioner/csi-hostpathplugin-9nmnq" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.743257 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e686de65-3d86-4826-9f0a-4132c81e9ed2-metrics-certs\") pod \"router-default-5444994796-h9kpk\" (UID: \"e686de65-3d86-4826-9f0a-4132c81e9ed2\") " pod="openshift-ingress/router-default-5444994796-h9kpk" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.743275 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cbnx\" (UniqueName: \"kubernetes.io/projected/e686de65-3d86-4826-9f0a-4132c81e9ed2-kube-api-access-2cbnx\") pod \"router-default-5444994796-h9kpk\" (UID: \"e686de65-3d86-4826-9f0a-4132c81e9ed2\") " pod="openshift-ingress/router-default-5444994796-h9kpk" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.743293 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2ea695f7-4e9e-4b12-8e0c-22431291576d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7hstm\" (UID: \"2ea695f7-4e9e-4b12-8e0c-22431291576d\") " pod="openshift-marketplace/marketplace-operator-79b997595-7hstm" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.743320 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/4d793912-71b9-4100-ab1a-e24a0d10dd7e-certs\") pod \"machine-config-server-vq52n\" (UID: \"4d793912-71b9-4100-ab1a-e24a0d10dd7e\") " pod="openshift-machine-config-operator/machine-config-server-vq52n" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.743338 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/12ddf317-c034-4154-9966-c8b502ec0d6c-signing-key\") pod \"service-ca-9c57cc56f-pp9bw\" (UID: \"12ddf317-c034-4154-9966-c8b502ec0d6c\") " pod="openshift-service-ca/service-ca-9c57cc56f-pp9bw" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.743355 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/12ddf317-c034-4154-9966-c8b502ec0d6c-signing-cabundle\") pod \"service-ca-9c57cc56f-pp9bw\" (UID: \"12ddf317-c034-4154-9966-c8b502ec0d6c\") " pod="openshift-service-ca/service-ca-9c57cc56f-pp9bw" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.743374 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0726d810-1988-4eb3-9077-e8892670dc8c-metrics-tls\") pod \"ingress-operator-5b745b69d9-gvkmx\" (UID: \"0726d810-1988-4eb3-9077-e8892670dc8c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-gvkmx" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.743394 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f99594fb-adb1-469b-960f-2a56dc3427bd-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-csmdk\" (UID: \"f99594fb-adb1-469b-960f-2a56dc3427bd\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-csmdk" Dec 01 08:42:40 crc kubenswrapper[4873]: E1201 08:42:40.743639 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:41.243618973 +0000 UTC m=+137.145727512 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.744415 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e686de65-3d86-4826-9f0a-4132c81e9ed2-service-ca-bundle\") pod \"router-default-5444994796-h9kpk\" (UID: \"e686de65-3d86-4826-9f0a-4132c81e9ed2\") " pod="openshift-ingress/router-default-5444994796-h9kpk" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.744984 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/04b9f090-d8f6-4122-bbff-af681fcc4f42-csi-data-dir\") pod \"csi-hostpathplugin-9nmnq\" (UID: \"04b9f090-d8f6-4122-bbff-af681fcc4f42\") " pod="hostpath-provisioner/csi-hostpathplugin-9nmnq" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.745857 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/83688ccb-ace0-4b0a-b584-e650671f5686-config-volume\") pod \"collect-profiles-29409630-lnw2h\" (UID: \"83688ccb-ace0-4b0a-b584-e650671f5686\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-lnw2h" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.747093 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9a6d7d20-0eea-42ae-92e0-91b1b555bf6f-config-volume\") pod \"dns-default-crgj6\" (UID: \"9a6d7d20-0eea-42ae-92e0-91b1b555bf6f\") " pod="openshift-dns/dns-default-crgj6" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.747869 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/04b9f090-d8f6-4122-bbff-af681fcc4f42-registration-dir\") pod \"csi-hostpathplugin-9nmnq\" (UID: \"04b9f090-d8f6-4122-bbff-af681fcc4f42\") " pod="hostpath-provisioner/csi-hostpathplugin-9nmnq" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.748675 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/812f722d-c159-40bb-8901-bac36d182280-config\") pod \"service-ca-operator-777779d784-x9tsn\" (UID: \"812f722d-c159-40bb-8901-bac36d182280\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-x9tsn" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.750142 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/04b9f090-d8f6-4122-bbff-af681fcc4f42-plugins-dir\") pod \"csi-hostpathplugin-9nmnq\" (UID: \"04b9f090-d8f6-4122-bbff-af681fcc4f42\") " pod="hostpath-provisioner/csi-hostpathplugin-9nmnq" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.750935 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/12ddf317-c034-4154-9966-c8b502ec0d6c-signing-cabundle\") pod \"service-ca-9c57cc56f-pp9bw\" (UID: \"12ddf317-c034-4154-9966-c8b502ec0d6c\") " pod="openshift-service-ca/service-ca-9c57cc56f-pp9bw" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.751922 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgbxj\" (UniqueName: \"kubernetes.io/projected/16511918-7dea-4667-9119-07f0da1b9a84-kube-api-access-wgbxj\") pod \"console-operator-58897d9998-5bhp6\" (UID: \"16511918-7dea-4667-9119-07f0da1b9a84\") " pod="openshift-console-operator/console-operator-58897d9998-5bhp6" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.754069 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/23e5ea75-df3a-4a09-87f4-542f0063f4aa-cert\") pod \"ingress-canary-pwc6t\" (UID: \"23e5ea75-df3a-4a09-87f4-542f0063f4aa\") " pod="openshift-ingress-canary/ingress-canary-pwc6t" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.754903 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7aee65fe-0d01-4db7-95ee-131e6ead9071-auth-proxy-config\") pod \"machine-config-operator-74547568cd-s5ct4\" (UID: \"7aee65fe-0d01-4db7-95ee-131e6ead9071\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s5ct4" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.755119 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/04b9f090-d8f6-4122-bbff-af681fcc4f42-socket-dir\") pod \"csi-hostpathplugin-9nmnq\" (UID: \"04b9f090-d8f6-4122-bbff-af681fcc4f42\") " pod="hostpath-provisioner/csi-hostpathplugin-9nmnq" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.755707 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/4d793912-71b9-4100-ab1a-e24a0d10dd7e-certs\") pod \"machine-config-server-vq52n\" (UID: \"4d793912-71b9-4100-ab1a-e24a0d10dd7e\") " pod="openshift-machine-config-operator/machine-config-server-vq52n" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.756294 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/83688ccb-ace0-4b0a-b584-e650671f5686-secret-volume\") pod \"collect-profiles-29409630-lnw2h\" (UID: \"83688ccb-ace0-4b0a-b584-e650671f5686\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-lnw2h" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.756783 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2ea695f7-4e9e-4b12-8e0c-22431291576d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7hstm\" (UID: \"2ea695f7-4e9e-4b12-8e0c-22431291576d\") " pod="openshift-marketplace/marketplace-operator-79b997595-7hstm" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.758816 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/04b9f090-d8f6-4122-bbff-af681fcc4f42-mountpoint-dir\") pod \"csi-hostpathplugin-9nmnq\" (UID: \"04b9f090-d8f6-4122-bbff-af681fcc4f42\") " pod="hostpath-provisioner/csi-hostpathplugin-9nmnq" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.759100 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb6d3b33-6120-40aa-8c4d-01f186d1ab17-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-xdpgx\" (UID: \"fb6d3b33-6120-40aa-8c4d-01f186d1ab17\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xdpgx" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.759180 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/e686de65-3d86-4826-9f0a-4132c81e9ed2-stats-auth\") pod \"router-default-5444994796-h9kpk\" (UID: \"e686de65-3d86-4826-9f0a-4132c81e9ed2\") " pod="openshift-ingress/router-default-5444994796-h9kpk" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.759728 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fb6d3b33-6120-40aa-8c4d-01f186d1ab17-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-xdpgx\" (UID: \"fb6d3b33-6120-40aa-8c4d-01f186d1ab17\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xdpgx" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.759944 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f99594fb-adb1-469b-960f-2a56dc3427bd-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-csmdk\" (UID: \"f99594fb-adb1-469b-960f-2a56dc3427bd\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-csmdk" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.760171 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3dc1fc6b-9c79-4a78-af19-d49195bc94be-srv-cert\") pod \"catalog-operator-68c6474976-whl8s\" (UID: \"3dc1fc6b-9c79-4a78-af19-d49195bc94be\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-whl8s" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.760563 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/12ddf317-c034-4154-9966-c8b502ec0d6c-signing-key\") pod \"service-ca-9c57cc56f-pp9bw\" (UID: \"12ddf317-c034-4154-9966-c8b502ec0d6c\") " pod="openshift-service-ca/service-ca-9c57cc56f-pp9bw" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.761148 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e3e97e3b-f082-4093-b183-fcd5ac857c07-srv-cert\") pod \"olm-operator-6b444d44fb-xrgx6\" (UID: \"e3e97e3b-f082-4093-b183-fcd5ac857c07\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xrgx6" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.761531 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/90dd16fd-d069-4007-8c82-4811cb343380-tmpfs\") pod \"packageserver-d55dfcdfc-f5rc8\" (UID: \"90dd16fd-d069-4007-8c82-4811cb343380\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f5rc8" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.761623 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f99594fb-adb1-469b-960f-2a56dc3427bd-config\") pod \"kube-apiserver-operator-766d6c64bb-csmdk\" (UID: \"f99594fb-adb1-469b-960f-2a56dc3427bd\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-csmdk" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.762238 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/4d793912-71b9-4100-ab1a-e24a0d10dd7e-node-bootstrap-token\") pod \"machine-config-server-vq52n\" (UID: \"4d793912-71b9-4100-ab1a-e24a0d10dd7e\") " pod="openshift-machine-config-operator/machine-config-server-vq52n" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.762535 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2ea695f7-4e9e-4b12-8e0c-22431291576d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7hstm\" (UID: \"2ea695f7-4e9e-4b12-8e0c-22431291576d\") " pod="openshift-marketplace/marketplace-operator-79b997595-7hstm" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.762849 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7aee65fe-0d01-4db7-95ee-131e6ead9071-images\") pod \"machine-config-operator-74547568cd-s5ct4\" (UID: \"7aee65fe-0d01-4db7-95ee-131e6ead9071\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s5ct4" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.762972 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e3e97e3b-f082-4093-b183-fcd5ac857c07-profile-collector-cert\") pod \"olm-operator-6b444d44fb-xrgx6\" (UID: \"e3e97e3b-f082-4093-b183-fcd5ac857c07\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xrgx6" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.763718 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/67e07bc1-ce0f-453c-83ba-6f8cc36ed889-metrics-tls\") pod \"dns-operator-744455d44c-f5kts\" (UID: \"67e07bc1-ce0f-453c-83ba-6f8cc36ed889\") " pod="openshift-dns-operator/dns-operator-744455d44c-f5kts" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.764307 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/2a86c401-c7dd-4b95-b267-8c58e5728e6a-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-mvczl\" (UID: \"2a86c401-c7dd-4b95-b267-8c58e5728e6a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mvczl" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.765059 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/90dd16fd-d069-4007-8c82-4811cb343380-apiservice-cert\") pod \"packageserver-d55dfcdfc-f5rc8\" (UID: \"90dd16fd-d069-4007-8c82-4811cb343380\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f5rc8" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.765535 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e686de65-3d86-4826-9f0a-4132c81e9ed2-metrics-certs\") pod \"router-default-5444994796-h9kpk\" (UID: \"e686de65-3d86-4826-9f0a-4132c81e9ed2\") " pod="openshift-ingress/router-default-5444994796-h9kpk" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.766007 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3dc1fc6b-9c79-4a78-af19-d49195bc94be-profile-collector-cert\") pod \"catalog-operator-68c6474976-whl8s\" (UID: \"3dc1fc6b-9c79-4a78-af19-d49195bc94be\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-whl8s" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.766153 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9th6b\" (UniqueName: \"kubernetes.io/projected/0d0699e2-c1b3-48a3-8d17-5d89424b4bb0-kube-api-access-9th6b\") pod \"machine-approver-56656f9798-wq2zz\" (UID: \"0d0699e2-c1b3-48a3-8d17-5d89424b4bb0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wq2zz" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.769646 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0726d810-1988-4eb3-9077-e8892670dc8c-metrics-tls\") pod \"ingress-operator-5b745b69d9-gvkmx\" (UID: \"0726d810-1988-4eb3-9077-e8892670dc8c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-gvkmx" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.792320 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0726d810-1988-4eb3-9077-e8892670dc8c-trusted-ca\") pod \"ingress-operator-5b745b69d9-gvkmx\" (UID: \"0726d810-1988-4eb3-9077-e8892670dc8c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-gvkmx" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.794530 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/e686de65-3d86-4826-9f0a-4132c81e9ed2-default-certificate\") pod \"router-default-5444994796-h9kpk\" (UID: \"e686de65-3d86-4826-9f0a-4132c81e9ed2\") " pod="openshift-ingress/router-default-5444994796-h9kpk" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.797562 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmww7\" (UniqueName: \"kubernetes.io/projected/d0c705dd-9608-4b5e-8c91-2511d8427419-kube-api-access-cmww7\") pod \"apiserver-76f77b778f-w8xn2\" (UID: \"d0c705dd-9608-4b5e-8c91-2511d8427419\") " pod="openshift-apiserver/apiserver-76f77b778f-w8xn2" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.815803 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qwvk\" (UniqueName: \"kubernetes.io/projected/83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5-kube-api-access-4qwvk\") pod \"route-controller-manager-6576b87f9c-q4wmp\" (UID: \"83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q4wmp" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.829083 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-w8xn2" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.837665 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7q6cm\" (UniqueName: \"kubernetes.io/projected/f2208bec-45d1-45f0-8700-1218abfa7642-kube-api-access-7q6cm\") pod \"cluster-image-registry-operator-dc59b4c8b-ptnr8\" (UID: \"f2208bec-45d1-45f0-8700-1218abfa7642\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ptnr8" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.844583 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:40 crc kubenswrapper[4873]: E1201 08:42:40.845234 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:41.345222203 +0000 UTC m=+137.247330732 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.847720 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7aee65fe-0d01-4db7-95ee-131e6ead9071-proxy-tls\") pod \"machine-config-operator-74547568cd-s5ct4\" (UID: \"7aee65fe-0d01-4db7-95ee-131e6ead9071\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s5ct4" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.852540 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/90dd16fd-d069-4007-8c82-4811cb343380-webhook-cert\") pod \"packageserver-d55dfcdfc-f5rc8\" (UID: \"90dd16fd-d069-4007-8c82-4811cb343380\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f5rc8" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.852651 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/812f722d-c159-40bb-8901-bac36d182280-serving-cert\") pod \"service-ca-operator-777779d784-x9tsn\" (UID: \"812f722d-c159-40bb-8901-bac36d182280\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-x9tsn" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.853246 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9a6d7d20-0eea-42ae-92e0-91b1b555bf6f-metrics-tls\") pod \"dns-default-crgj6\" (UID: \"9a6d7d20-0eea-42ae-92e0-91b1b555bf6f\") " pod="openshift-dns/dns-default-crgj6" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.854918 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/952ae5ba-156f-41a7-950a-1783e2e69f94-bound-sa-token\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.867181 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7dcf8" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.873377 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gscbx" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.883268 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z78ht\" (UniqueName: \"kubernetes.io/projected/17e954cb-2d58-4616-bd5c-e813644d411d-kube-api-access-z78ht\") pod \"multus-admission-controller-857f4d67dd-qvzhc\" (UID: \"17e954cb-2d58-4616-bd5c-e813644d411d\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-qvzhc" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.883334 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-5bhp6" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.929565 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wq2zz" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.946041 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:40 crc kubenswrapper[4873]: E1201 08:42:40.946759 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:41.446733259 +0000 UTC m=+137.348841798 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.958455 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwdlq\" (UniqueName: \"kubernetes.io/projected/757ee90a-18e3-4ccb-b03c-43e90bf8cb34-kube-api-access-gwdlq\") pod \"etcd-operator-b45778765-xt48p\" (UID: \"757ee90a-18e3-4ccb-b03c-43e90bf8cb34\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xt48p" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.973211 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ptnr8" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.980273 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-292nl" Dec 01 08:42:40 crc kubenswrapper[4873]: I1201 08:42:40.994415 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-qvzhc" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.066445 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72gzp\" (UniqueName: \"kubernetes.io/projected/83688ccb-ace0-4b0a-b584-e650671f5686-kube-api-access-72gzp\") pod \"collect-profiles-29409630-lnw2h\" (UID: \"83688ccb-ace0-4b0a-b584-e650671f5686\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-lnw2h" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.072266 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.072329 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q4wmp" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.072461 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlc6g\" (UniqueName: \"kubernetes.io/projected/952ae5ba-156f-41a7-950a-1783e2e69f94-kube-api-access-qlc6g\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:41 crc kubenswrapper[4873]: E1201 08:42:41.073062 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:41.573045893 +0000 UTC m=+137.475154442 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.074460 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwvvt\" (UniqueName: \"kubernetes.io/projected/10d7dabb-aa6f-4f2b-acf9-b02dc369d8be-kube-api-access-rwvvt\") pod \"cluster-samples-operator-665b6dd947-nd5kp\" (UID: \"10d7dabb-aa6f-4f2b-acf9-b02dc369d8be\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-nd5kp" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.086097 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xv9q\" (UniqueName: \"kubernetes.io/projected/2404e1f8-203a-4055-8f66-21d8867e7a92-kube-api-access-4xv9q\") pod \"oauth-openshift-558db77b4-sg6bg\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.088694 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c92cq\" (UniqueName: \"kubernetes.io/projected/ae767d9a-3df4-413a-bc25-8fc819c7088e-kube-api-access-c92cq\") pod \"openshift-apiserver-operator-796bbdcf4f-xfb2s\" (UID: \"ae767d9a-3df4-413a-bc25-8fc819c7088e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xfb2s" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.089755 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25wjr\" (UniqueName: \"kubernetes.io/projected/4af31a76-3090-497d-b39f-965c27a27a1c-kube-api-access-25wjr\") pod \"machine-config-controller-84d6567774-gxhcq\" (UID: \"4af31a76-3090-497d-b39f-965c27a27a1c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gxhcq" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.091034 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f99594fb-adb1-469b-960f-2a56dc3427bd-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-csmdk\" (UID: \"f99594fb-adb1-469b-960f-2a56dc3427bd\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-csmdk" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.098995 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-lnw2h" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.104661 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dcn4\" (UniqueName: \"kubernetes.io/projected/67e07bc1-ce0f-453c-83ba-6f8cc36ed889-kube-api-access-9dcn4\") pod \"dns-operator-744455d44c-f5kts\" (UID: \"67e07bc1-ce0f-453c-83ba-6f8cc36ed889\") " pod="openshift-dns-operator/dns-operator-744455d44c-f5kts" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.108940 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dd6b\" (UniqueName: \"kubernetes.io/projected/2a86c401-c7dd-4b95-b267-8c58e5728e6a-kube-api-access-8dd6b\") pod \"package-server-manager-789f6589d5-mvczl\" (UID: \"2a86c401-c7dd-4b95-b267-8c58e5728e6a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mvczl" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.115533 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8xhk\" (UniqueName: \"kubernetes.io/projected/e3e97e3b-f082-4093-b183-fcd5ac857c07-kube-api-access-n8xhk\") pod \"olm-operator-6b444d44fb-xrgx6\" (UID: \"e3e97e3b-f082-4093-b183-fcd5ac857c07\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xrgx6" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.134877 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mvczl" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.174268 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:41 crc kubenswrapper[4873]: E1201 08:42:41.174691 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:41.674670264 +0000 UTC m=+137.576778803 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.174803 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-xt48p" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.174954 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cbnx\" (UniqueName: \"kubernetes.io/projected/e686de65-3d86-4826-9f0a-4132c81e9ed2-kube-api-access-2cbnx\") pod \"router-default-5444994796-h9kpk\" (UID: \"e686de65-3d86-4826-9f0a-4132c81e9ed2\") " pod="openshift-ingress/router-default-5444994796-h9kpk" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.175828 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnvrw\" (UniqueName: \"kubernetes.io/projected/7aee65fe-0d01-4db7-95ee-131e6ead9071-kube-api-access-lnvrw\") pod \"machine-config-operator-74547568cd-s5ct4\" (UID: \"7aee65fe-0d01-4db7-95ee-131e6ead9071\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s5ct4" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.255437 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-nd5kp" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.256040 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.257831 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xfb2s" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.278414 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkwmv\" (UniqueName: \"kubernetes.io/projected/3dc1fc6b-9c79-4a78-af19-d49195bc94be-kube-api-access-mkwmv\") pod \"catalog-operator-68c6474976-whl8s\" (UID: \"3dc1fc6b-9c79-4a78-af19-d49195bc94be\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-whl8s" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.281211 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:41 crc kubenswrapper[4873]: E1201 08:42:41.281735 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:41.78171766 +0000 UTC m=+137.683826199 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.282831 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpfsm\" (UniqueName: \"kubernetes.io/projected/90dd16fd-d069-4007-8c82-4811cb343380-kube-api-access-zpfsm\") pod \"packageserver-d55dfcdfc-f5rc8\" (UID: \"90dd16fd-d069-4007-8c82-4811cb343380\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f5rc8" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.283915 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7n4g\" (UniqueName: \"kubernetes.io/projected/0726d810-1988-4eb3-9077-e8892670dc8c-kube-api-access-z7n4g\") pod \"ingress-operator-5b745b69d9-gvkmx\" (UID: \"0726d810-1988-4eb3-9077-e8892670dc8c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-gvkmx" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.295055 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fttq4\" (UniqueName: \"kubernetes.io/projected/04b9f090-d8f6-4122-bbff-af681fcc4f42-kube-api-access-fttq4\") pod \"csi-hostpathplugin-9nmnq\" (UID: \"04b9f090-d8f6-4122-bbff-af681fcc4f42\") " pod="hostpath-provisioner/csi-hostpathplugin-9nmnq" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.299440 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25z7h\" (UniqueName: \"kubernetes.io/projected/dfbcf0e0-e5a1-41e0-813c-008f355a1f0a-kube-api-access-25z7h\") pod \"migrator-59844c95c7-l9mq8\" (UID: \"dfbcf0e0-e5a1-41e0-813c-008f355a1f0a\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-l9mq8" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.306602 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gxhcq" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.306840 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7hln\" (UniqueName: \"kubernetes.io/projected/812f722d-c159-40bb-8901-bac36d182280-kube-api-access-w7hln\") pod \"service-ca-operator-777779d784-x9tsn\" (UID: \"812f722d-c159-40bb-8901-bac36d182280\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-x9tsn" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.326156 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-h9kpk" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.334455 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72t8c\" (UniqueName: \"kubernetes.io/projected/2ea695f7-4e9e-4b12-8e0c-22431291576d-kube-api-access-72t8c\") pod \"marketplace-operator-79b997595-7hstm\" (UID: \"2ea695f7-4e9e-4b12-8e0c-22431291576d\") " pod="openshift-marketplace/marketplace-operator-79b997595-7hstm" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.335238 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bq77\" (UniqueName: \"kubernetes.io/projected/23e5ea75-df3a-4a09-87f4-542f0063f4aa-kube-api-access-6bq77\") pod \"ingress-canary-pwc6t\" (UID: \"23e5ea75-df3a-4a09-87f4-542f0063f4aa\") " pod="openshift-ingress-canary/ingress-canary-pwc6t" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.335562 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-f5kts" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.342179 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-csmdk" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.358361 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-l9mq8" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.367794 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-x9tsn" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.369702 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrqmg\" (UniqueName: \"kubernetes.io/projected/12ddf317-c034-4154-9966-c8b502ec0d6c-kube-api-access-nrqmg\") pod \"service-ca-9c57cc56f-pp9bw\" (UID: \"12ddf317-c034-4154-9966-c8b502ec0d6c\") " pod="openshift-service-ca/service-ca-9c57cc56f-pp9bw" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.375596 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s5ct4" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.375857 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cp6l\" (UniqueName: \"kubernetes.io/projected/9a6d7d20-0eea-42ae-92e0-91b1b555bf6f-kube-api-access-8cp6l\") pod \"dns-default-crgj6\" (UID: \"9a6d7d20-0eea-42ae-92e0-91b1b555bf6f\") " pod="openshift-dns/dns-default-crgj6" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.384538 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-whl8s" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.384854 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0726d810-1988-4eb3-9077-e8892670dc8c-bound-sa-token\") pod \"ingress-operator-5b745b69d9-gvkmx\" (UID: \"0726d810-1988-4eb3-9077-e8892670dc8c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-gvkmx" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.384967 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:41 crc kubenswrapper[4873]: E1201 08:42:41.385109 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:41.885078083 +0000 UTC m=+137.787186622 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.385300 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:41 crc kubenswrapper[4873]: E1201 08:42:41.385684 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:41.885675537 +0000 UTC m=+137.787784076 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.399739 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nm59\" (UniqueName: \"kubernetes.io/projected/4d793912-71b9-4100-ab1a-e24a0d10dd7e-kube-api-access-8nm59\") pod \"machine-config-server-vq52n\" (UID: \"4d793912-71b9-4100-ab1a-e24a0d10dd7e\") " pod="openshift-machine-config-operator/machine-config-server-vq52n" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.409420 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xrgx6" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.455341 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-crgj6" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.456543 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-vq52n" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.456779 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f5rc8" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.456989 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-pp9bw" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.486639 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.486997 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-pwc6t" Dec 01 08:42:41 crc kubenswrapper[4873]: E1201 08:42:41.487495 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:41.987473685 +0000 UTC m=+137.889582224 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.487984 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-9nmnq" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.488423 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldzx7\" (UniqueName: \"kubernetes.io/projected/fb6d3b33-6120-40aa-8c4d-01f186d1ab17-kube-api-access-ldzx7\") pod \"kube-storage-version-migrator-operator-b67b599dd-xdpgx\" (UID: \"fb6d3b33-6120-40aa-8c4d-01f186d1ab17\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xdpgx" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.519543 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wq2zz" event={"ID":"0d0699e2-c1b3-48a3-8d17-5d89424b4bb0","Type":"ContainerStarted","Data":"efee5b22ebfc6d9d96c77d95ed711ea3c01413b74f3f14495f60890da9367be4"} Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.524683 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tdwnv" event={"ID":"54e431c7-f38b-4538-a9f6-69d7a8995dce","Type":"ContainerStarted","Data":"64ee772ec468ed5794cf787708446f045f7b08feb36abe0e842888af9aa67368"} Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.525601 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-z6qdr" event={"ID":"1ea01a0b-ae00-469d-a56d-0103ac067d1b","Type":"ContainerStarted","Data":"16bb50439e1d2fdd0e19027a0e0ec9be62c94d7844539282830617fb19d5ba1f"} Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.529457 4873 generic.go:334] "Generic (PLEG): container finished" podID="af90c16f-64c2-425f-843e-dd3a96fa9533" containerID="a75cc271b3ed6df8fc621714d2ae404799b34383c4527b1e3a8799ad079fa379" exitCode=0 Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.529592 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cl5cd" event={"ID":"af90c16f-64c2-425f-843e-dd3a96fa9533","Type":"ContainerDied","Data":"a75cc271b3ed6df8fc621714d2ae404799b34383c4527b1e3a8799ad079fa379"} Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.536320 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xg89d" event={"ID":"f4709124-2f04-4d3c-abb7-30a5d0925b6c","Type":"ContainerStarted","Data":"fa355581fbd6eaa0568860526fd306d486820947bed61e80397f3369f5b22270"} Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.536371 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xg89d" event={"ID":"f4709124-2f04-4d3c-abb7-30a5d0925b6c","Type":"ContainerStarted","Data":"4344d13f31299182ba5be41b4756c51567064131081dd1e9755a069b98e39a61"} Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.540538 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-76drh" event={"ID":"dc0adee4-65bf-484b-8f55-3c83022cae82","Type":"ContainerStarted","Data":"1499419e5b57727620327cee76c858d95e80836b69c2d2fbf51c218d1590b169"} Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.545225 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-52pjn" event={"ID":"fb8fcfca-0933-4a74-85de-859fda0153d6","Type":"ContainerStarted","Data":"2ec3c17fbc2bc858fa0ec24b3fb1301c9519263e4931db1f9412e435d85d3f25"} Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.576899 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-kw4m8" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.592214 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:41 crc kubenswrapper[4873]: E1201 08:42:41.594106 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:42.094092562 +0000 UTC m=+137.996201101 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.622003 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7hstm" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.652654 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-gvkmx" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.658617 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xdpgx" Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.662595 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-w8xn2"] Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.703889 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:41 crc kubenswrapper[4873]: E1201 08:42:41.704503 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:42.204482717 +0000 UTC m=+138.106591256 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.805872 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:41 crc kubenswrapper[4873]: E1201 08:42:41.806273 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:42.306258294 +0000 UTC m=+138.208366833 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:41 crc kubenswrapper[4873]: I1201 08:42:41.907480 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:41 crc kubenswrapper[4873]: E1201 08:42:41.908473 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:42.408454218 +0000 UTC m=+138.310562757 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:42 crc kubenswrapper[4873]: I1201 08:42:42.009589 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:42 crc kubenswrapper[4873]: E1201 08:42:42.010233 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:42.510219015 +0000 UTC m=+138.412327554 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:42 crc kubenswrapper[4873]: I1201 08:42:42.018969 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-5bhp6"] Dec 01 08:42:42 crc kubenswrapper[4873]: I1201 08:42:42.110390 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:42 crc kubenswrapper[4873]: E1201 08:42:42.110987 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:42.610964149 +0000 UTC m=+138.513072688 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:42 crc kubenswrapper[4873]: I1201 08:42:42.211870 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:42 crc kubenswrapper[4873]: E1201 08:42:42.212326 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:42.712307919 +0000 UTC m=+138.614416448 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:42 crc kubenswrapper[4873]: I1201 08:42:42.313060 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:42 crc kubenswrapper[4873]: E1201 08:42:42.313538 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:42.813511202 +0000 UTC m=+138.715619751 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:42 crc kubenswrapper[4873]: I1201 08:42:42.313981 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:42 crc kubenswrapper[4873]: E1201 08:42:42.314501 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:42.814489313 +0000 UTC m=+138.716597852 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:42 crc kubenswrapper[4873]: I1201 08:42:42.415781 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:42 crc kubenswrapper[4873]: E1201 08:42:42.416234 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:42.916197587 +0000 UTC m=+138.818306126 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:42 crc kubenswrapper[4873]: I1201 08:42:42.416581 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:42 crc kubenswrapper[4873]: E1201 08:42:42.416974 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:42.916961739 +0000 UTC m=+138.819070278 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:42 crc kubenswrapper[4873]: I1201 08:42:42.517572 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:42 crc kubenswrapper[4873]: E1201 08:42:42.517705 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:43.017679132 +0000 UTC m=+138.919787671 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:42 crc kubenswrapper[4873]: I1201 08:42:42.517797 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:42 crc kubenswrapper[4873]: E1201 08:42:42.518084 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:43.018077819 +0000 UTC m=+138.920186358 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:42 crc kubenswrapper[4873]: I1201 08:42:42.632333 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:42 crc kubenswrapper[4873]: E1201 08:42:42.633221 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:43.133192539 +0000 UTC m=+139.035301078 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:42 crc kubenswrapper[4873]: I1201 08:42:42.717455 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-h9kpk" event={"ID":"e686de65-3d86-4826-9f0a-4132c81e9ed2","Type":"ContainerStarted","Data":"7aa825320a283c121432bf70421a83997f689c3bebae8118d7c50cc01d75521b"} Dec 01 08:42:42 crc kubenswrapper[4873]: I1201 08:42:42.734728 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:42 crc kubenswrapper[4873]: E1201 08:42:42.735278 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:43.235259948 +0000 UTC m=+139.137368487 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:42 crc kubenswrapper[4873]: I1201 08:42:42.796758 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-5bhp6" event={"ID":"16511918-7dea-4667-9119-07f0da1b9a84","Type":"ContainerStarted","Data":"3ea8d038353496b5aada71638c3bd4f1a64e3c63d0834e48b8cd6736fe3fa603"} Dec 01 08:42:42 crc kubenswrapper[4873]: I1201 08:42:42.819488 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tdwnv" event={"ID":"54e431c7-f38b-4538-a9f6-69d7a8995dce","Type":"ContainerStarted","Data":"232a0ecb0ae6c1cb772c535232df8a403da2fa8551ab6df583f3b7627aff8311"} Dec 01 08:42:42 crc kubenswrapper[4873]: I1201 08:42:42.840289 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:42 crc kubenswrapper[4873]: E1201 08:42:42.840534 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:43.34051539 +0000 UTC m=+139.242623929 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:42 crc kubenswrapper[4873]: I1201 08:42:42.840861 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-52pjn" podStartSLOduration=119.840849063 podStartE2EDuration="1m59.840849063s" podCreationTimestamp="2025-12-01 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:42.840613344 +0000 UTC m=+138.742721883" watchObservedRunningTime="2025-12-01 08:42:42.840849063 +0000 UTC m=+138.742957602" Dec 01 08:42:42 crc kubenswrapper[4873]: I1201 08:42:42.885878 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wq2zz" event={"ID":"0d0699e2-c1b3-48a3-8d17-5d89424b4bb0","Type":"ContainerStarted","Data":"b8f55cff9dbe452201b4432b3c42bb69db48dbe6c6c5e5261cd9bac8e44a7841"} Dec 01 08:42:42 crc kubenswrapper[4873]: I1201 08:42:42.887631 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-z6qdr" event={"ID":"1ea01a0b-ae00-469d-a56d-0103ac067d1b","Type":"ContainerStarted","Data":"aef0188864774094d7796227107ad68434c3d47ed880ea7ca3c3de05efb10b12"} Dec 01 08:42:42 crc kubenswrapper[4873]: I1201 08:42:42.929953 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-kw4m8" podStartSLOduration=119.929930055 podStartE2EDuration="1m59.929930055s" podCreationTimestamp="2025-12-01 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:42.928702444 +0000 UTC m=+138.830810983" watchObservedRunningTime="2025-12-01 08:42:42.929930055 +0000 UTC m=+138.832038594" Dec 01 08:42:42 crc kubenswrapper[4873]: I1201 08:42:42.950900 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:42 crc kubenswrapper[4873]: E1201 08:42:42.951344 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:43.451325711 +0000 UTC m=+139.353434250 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:42 crc kubenswrapper[4873]: I1201 08:42:42.952250 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lcqht" event={"ID":"01924913-40f0-4502-b151-36079ee0ea8c","Type":"ContainerStarted","Data":"ef79b6fc510381c197f518e84cd13c8c155c26c20a07b96e0a966084895f7bc8"} Dec 01 08:42:42 crc kubenswrapper[4873]: I1201 08:42:42.953754 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-q4wmp"] Dec 01 08:42:42 crc kubenswrapper[4873]: I1201 08:42:42.964892 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-76drh" event={"ID":"dc0adee4-65bf-484b-8f55-3c83022cae82","Type":"ContainerStarted","Data":"c1f67bcaf957ca4bf14a6e56e284d599b4ab2c87af4c805b91befb72c1394c72"} Dec 01 08:42:42 crc kubenswrapper[4873]: I1201 08:42:42.966264 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-76drh" Dec 01 08:42:42 crc kubenswrapper[4873]: I1201 08:42:42.971406 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-w8xn2" event={"ID":"d0c705dd-9608-4b5e-8c91-2511d8427419","Type":"ContainerStarted","Data":"2851437e8ae176b03df95a5f5a26b71468e493f9768e80f57d0002ff358b3244"} Dec 01 08:42:42 crc kubenswrapper[4873]: I1201 08:42:42.985375 4873 patch_prober.go:28] interesting pod/downloads-7954f5f757-76drh container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Dec 01 08:42:42 crc kubenswrapper[4873]: I1201 08:42:42.985436 4873 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-76drh" podUID="dc0adee4-65bf-484b-8f55-3c83022cae82" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Dec 01 08:42:42 crc kubenswrapper[4873]: I1201 08:42:42.989235 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-vq52n" event={"ID":"4d793912-71b9-4100-ab1a-e24a0d10dd7e","Type":"ContainerStarted","Data":"16dba51f748bd3bcb7a0a8ff7c6a972d2ee1b2e0751089fa6ac3275676b9fe10"} Dec 01 08:42:43 crc kubenswrapper[4873]: I1201 08:42:43.052409 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:43 crc kubenswrapper[4873]: E1201 08:42:43.053974 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:43.553950334 +0000 UTC m=+139.456058873 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:43 crc kubenswrapper[4873]: I1201 08:42:43.144120 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xg89d" podStartSLOduration=120.144098409 podStartE2EDuration="2m0.144098409s" podCreationTimestamp="2025-12-01 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:43.052207621 +0000 UTC m=+138.954316160" watchObservedRunningTime="2025-12-01 08:42:43.144098409 +0000 UTC m=+139.046206948" Dec 01 08:42:43 crc kubenswrapper[4873]: I1201 08:42:43.171154 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:43 crc kubenswrapper[4873]: E1201 08:42:43.172943 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:43.672922013 +0000 UTC m=+139.575030642 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:43 crc kubenswrapper[4873]: I1201 08:42:43.303634 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:43 crc kubenswrapper[4873]: E1201 08:42:43.304141 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:43.804111729 +0000 UTC m=+139.706220268 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:43 crc kubenswrapper[4873]: I1201 08:42:43.304413 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:43 crc kubenswrapper[4873]: E1201 08:42:43.304986 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:43.804970715 +0000 UTC m=+139.707079254 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:43 crc kubenswrapper[4873]: I1201 08:42:43.407925 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:43 crc kubenswrapper[4873]: E1201 08:42:43.408704 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:43.908686412 +0000 UTC m=+139.810794951 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:43 crc kubenswrapper[4873]: I1201 08:42:43.490753 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-z6qdr" podStartSLOduration=120.490731042 podStartE2EDuration="2m0.490731042s" podCreationTimestamp="2025-12-01 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:43.488781311 +0000 UTC m=+139.390889880" watchObservedRunningTime="2025-12-01 08:42:43.490731042 +0000 UTC m=+139.392839581" Dec 01 08:42:43 crc kubenswrapper[4873]: I1201 08:42:43.505313 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-vq52n" podStartSLOduration=5.505289365 podStartE2EDuration="5.505289365s" podCreationTimestamp="2025-12-01 08:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:43.503105915 +0000 UTC m=+139.405214454" watchObservedRunningTime="2025-12-01 08:42:43.505289365 +0000 UTC m=+139.407397904" Dec 01 08:42:43 crc kubenswrapper[4873]: I1201 08:42:43.510858 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:43 crc kubenswrapper[4873]: E1201 08:42:43.511338 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:44.011324765 +0000 UTC m=+139.913433304 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:43 crc kubenswrapper[4873]: I1201 08:42:43.527125 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lcqht" podStartSLOduration=120.527106609 podStartE2EDuration="2m0.527106609s" podCreationTimestamp="2025-12-01 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:43.524551984 +0000 UTC m=+139.426660523" watchObservedRunningTime="2025-12-01 08:42:43.527106609 +0000 UTC m=+139.429215148" Dec 01 08:42:43 crc kubenswrapper[4873]: I1201 08:42:43.596056 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tdwnv" podStartSLOduration=120.596026335 podStartE2EDuration="2m0.596026335s" podCreationTimestamp="2025-12-01 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:43.592950068 +0000 UTC m=+139.495058607" watchObservedRunningTime="2025-12-01 08:42:43.596026335 +0000 UTC m=+139.498134894" Dec 01 08:42:43 crc kubenswrapper[4873]: I1201 08:42:43.613283 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:43 crc kubenswrapper[4873]: E1201 08:42:43.613971 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:44.113923607 +0000 UTC m=+140.016032146 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:43 crc kubenswrapper[4873]: I1201 08:42:43.621602 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-76drh" podStartSLOduration=120.62148174 podStartE2EDuration="2m0.62148174s" podCreationTimestamp="2025-12-01 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:43.615505832 +0000 UTC m=+139.517614371" watchObservedRunningTime="2025-12-01 08:42:43.62148174 +0000 UTC m=+139.523590279" Dec 01 08:42:43 crc kubenswrapper[4873]: I1201 08:42:43.715848 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:43 crc kubenswrapper[4873]: E1201 08:42:43.716363 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:44.216346921 +0000 UTC m=+140.118455460 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:43 crc kubenswrapper[4873]: I1201 08:42:43.817164 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:43 crc kubenswrapper[4873]: E1201 08:42:43.817407 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:44.317363847 +0000 UTC m=+140.219472386 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:43 crc kubenswrapper[4873]: I1201 08:42:43.817688 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:43 crc kubenswrapper[4873]: E1201 08:42:43.818068 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:44.318053635 +0000 UTC m=+140.220162174 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:43 crc kubenswrapper[4873]: I1201 08:42:43.920828 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:43 crc kubenswrapper[4873]: E1201 08:42:43.921326 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:44.421289173 +0000 UTC m=+140.323397712 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:43 crc kubenswrapper[4873]: I1201 08:42:43.921453 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:43 crc kubenswrapper[4873]: E1201 08:42:43.921840 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:44.421828066 +0000 UTC m=+140.323936605 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:44 crc kubenswrapper[4873]: I1201 08:42:44.030131 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:44 crc kubenswrapper[4873]: E1201 08:42:44.030492 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:44.530442376 +0000 UTC m=+140.432550915 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:44 crc kubenswrapper[4873]: I1201 08:42:44.037946 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:44 crc kubenswrapper[4873]: E1201 08:42:44.042168 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:44.54212426 +0000 UTC m=+140.444232969 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:44 crc kubenswrapper[4873]: I1201 08:42:44.045394 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-vq52n" event={"ID":"4d793912-71b9-4100-ab1a-e24a0d10dd7e","Type":"ContainerStarted","Data":"8e0c0d5344a465af5c2018bae9d27193f4f89be81792070a8c08a3a40129b978"} Dec 01 08:42:44 crc kubenswrapper[4873]: I1201 08:42:44.053575 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-h9kpk" event={"ID":"e686de65-3d86-4826-9f0a-4132c81e9ed2","Type":"ContainerStarted","Data":"edbb3e7e35126d781e0452779c9ad85be0a50049881fd7b1757208f3f8f98e12"} Dec 01 08:42:44 crc kubenswrapper[4873]: I1201 08:42:44.069458 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cl5cd" event={"ID":"af90c16f-64c2-425f-843e-dd3a96fa9533","Type":"ContainerStarted","Data":"28374cd2ca8c9f612da3362b6de4fb95b159024a3ffd67d7cf2b6408536b02e7"} Dec 01 08:42:44 crc kubenswrapper[4873]: I1201 08:42:44.072381 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cl5cd" Dec 01 08:42:44 crc kubenswrapper[4873]: I1201 08:42:44.080032 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-5bhp6" event={"ID":"16511918-7dea-4667-9119-07f0da1b9a84","Type":"ContainerStarted","Data":"d156713833a266f079506dd9ab1fc026ba9e46e20cb81ed2fb0940ac3871ea4e"} Dec 01 08:42:44 crc kubenswrapper[4873]: I1201 08:42:44.081205 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-5bhp6" Dec 01 08:42:44 crc kubenswrapper[4873]: I1201 08:42:44.089995 4873 patch_prober.go:28] interesting pod/console-operator-58897d9998-5bhp6 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/readyz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Dec 01 08:42:44 crc kubenswrapper[4873]: I1201 08:42:44.090081 4873 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-5bhp6" podUID="16511918-7dea-4667-9119-07f0da1b9a84" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/readyz\": dial tcp 10.217.0.13:8443: connect: connection refused" Dec 01 08:42:44 crc kubenswrapper[4873]: I1201 08:42:44.091523 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wq2zz" event={"ID":"0d0699e2-c1b3-48a3-8d17-5d89424b4bb0","Type":"ContainerStarted","Data":"e25d1c7c900865520fed44dcfd953309911de8646cd1e043762fd6cb07b978a9"} Dec 01 08:42:44 crc kubenswrapper[4873]: I1201 08:42:44.112975 4873 generic.go:334] "Generic (PLEG): container finished" podID="d0c705dd-9608-4b5e-8c91-2511d8427419" containerID="c1ccb3dfdd645e0c69449294ae88249c6fc24dfe5140905002db536873bbf0c4" exitCode=0 Dec 01 08:42:44 crc kubenswrapper[4873]: I1201 08:42:44.113066 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-w8xn2" event={"ID":"d0c705dd-9608-4b5e-8c91-2511d8427419","Type":"ContainerDied","Data":"c1ccb3dfdd645e0c69449294ae88249c6fc24dfe5140905002db536873bbf0c4"} Dec 01 08:42:44 crc kubenswrapper[4873]: I1201 08:42:44.122579 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q4wmp" event={"ID":"83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5","Type":"ContainerStarted","Data":"9124d6fa4df2506e2a75d9fe09d4a4aff028c3761c7506cfa5cbda7f6911b228"} Dec 01 08:42:44 crc kubenswrapper[4873]: I1201 08:42:44.123733 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q4wmp" event={"ID":"83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5","Type":"ContainerStarted","Data":"cd5b64af3eea625553caa17357eda8565a7cef89341a8b134d32d289d053e807"} Dec 01 08:42:44 crc kubenswrapper[4873]: I1201 08:42:44.123856 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q4wmp" Dec 01 08:42:44 crc kubenswrapper[4873]: I1201 08:42:44.124374 4873 patch_prober.go:28] interesting pod/downloads-7954f5f757-76drh container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Dec 01 08:42:44 crc kubenswrapper[4873]: I1201 08:42:44.124919 4873 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-76drh" podUID="dc0adee4-65bf-484b-8f55-3c83022cae82" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Dec 01 08:42:44 crc kubenswrapper[4873]: I1201 08:42:44.132990 4873 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-q4wmp container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Dec 01 08:42:44 crc kubenswrapper[4873]: I1201 08:42:44.133078 4873 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q4wmp" podUID="83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" Dec 01 08:42:44 crc kubenswrapper[4873]: I1201 08:42:44.145132 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:44 crc kubenswrapper[4873]: E1201 08:42:44.145307 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:44.645275434 +0000 UTC m=+140.547383973 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:44 crc kubenswrapper[4873]: I1201 08:42:44.145660 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:44 crc kubenswrapper[4873]: E1201 08:42:44.149833 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:44.649815863 +0000 UTC m=+140.551924402 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:44 crc kubenswrapper[4873]: I1201 08:42:44.223219 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-h9kpk" podStartSLOduration=121.223189653 podStartE2EDuration="2m1.223189653s" podCreationTimestamp="2025-12-01 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:44.216334589 +0000 UTC m=+140.118443128" watchObservedRunningTime="2025-12-01 08:42:44.223189653 +0000 UTC m=+140.125298182" Dec 01 08:42:44 crc kubenswrapper[4873]: I1201 08:42:44.226068 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7dcf8"] Dec 01 08:42:44 crc kubenswrapper[4873]: I1201 08:42:44.251792 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:44 crc kubenswrapper[4873]: E1201 08:42:44.253514 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:44.753490639 +0000 UTC m=+140.655599178 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:44 crc kubenswrapper[4873]: I1201 08:42:44.277214 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cl5cd" podStartSLOduration=121.277189141 podStartE2EDuration="2m1.277189141s" podCreationTimestamp="2025-12-01 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:44.276242651 +0000 UTC m=+140.178351200" watchObservedRunningTime="2025-12-01 08:42:44.277189141 +0000 UTC m=+140.179297680" Dec 01 08:42:44 crc kubenswrapper[4873]: I1201 08:42:44.328082 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-h9kpk" Dec 01 08:42:44 crc kubenswrapper[4873]: I1201 08:42:44.354891 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q4wmp" podStartSLOduration=121.354869909 podStartE2EDuration="2m1.354869909s" podCreationTimestamp="2025-12-01 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:44.308850842 +0000 UTC m=+140.210959401" watchObservedRunningTime="2025-12-01 08:42:44.354869909 +0000 UTC m=+140.256978448" Dec 01 08:42:44 crc kubenswrapper[4873]: I1201 08:42:44.360890 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:44 crc kubenswrapper[4873]: E1201 08:42:44.361715 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:44.861698252 +0000 UTC m=+140.763806791 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:44 crc kubenswrapper[4873]: I1201 08:42:44.369095 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ptnr8"] Dec 01 08:42:44 crc kubenswrapper[4873]: I1201 08:42:44.374956 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-5bhp6" podStartSLOduration=121.374930091 podStartE2EDuration="2m1.374930091s" podCreationTimestamp="2025-12-01 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:44.351843664 +0000 UTC m=+140.253952203" watchObservedRunningTime="2025-12-01 08:42:44.374930091 +0000 UTC m=+140.277038630" Dec 01 08:42:44 crc kubenswrapper[4873]: I1201 08:42:44.387528 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gscbx"] Dec 01 08:42:44 crc kubenswrapper[4873]: I1201 08:42:44.388630 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wq2zz" podStartSLOduration=121.388607237 podStartE2EDuration="2m1.388607237s" podCreationTimestamp="2025-12-01 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:44.380296973 +0000 UTC m=+140.282405512" watchObservedRunningTime="2025-12-01 08:42:44.388607237 +0000 UTC m=+140.290715776" Dec 01 08:42:44 crc kubenswrapper[4873]: I1201 08:42:44.401437 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-qvzhc"] Dec 01 08:42:44 crc kubenswrapper[4873]: I1201 08:42:44.446727 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lcqht" Dec 01 08:42:44 crc kubenswrapper[4873]: I1201 08:42:44.447220 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lcqht" Dec 01 08:42:44 crc kubenswrapper[4873]: I1201 08:42:44.465220 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:44 crc kubenswrapper[4873]: E1201 08:42:44.465655 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:44.965630309 +0000 UTC m=+140.867738848 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:44 crc kubenswrapper[4873]: I1201 08:42:44.574849 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:44 crc kubenswrapper[4873]: E1201 08:42:44.576383 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:45.076359637 +0000 UTC m=+140.978468176 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:44 crc kubenswrapper[4873]: I1201 08:42:44.694639 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:44 crc kubenswrapper[4873]: E1201 08:42:44.694827 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:45.194797415 +0000 UTC m=+141.096905944 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:44 crc kubenswrapper[4873]: I1201 08:42:44.695677 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:44 crc kubenswrapper[4873]: E1201 08:42:44.696120 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:45.196108379 +0000 UTC m=+141.098216918 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:44 crc kubenswrapper[4873]: I1201 08:42:44.704568 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lcqht" Dec 01 08:42:44 crc kubenswrapper[4873]: I1201 08:42:44.797655 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:44 crc kubenswrapper[4873]: E1201 08:42:44.797867 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:45.297835964 +0000 UTC m=+141.199944503 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:44 crc kubenswrapper[4873]: I1201 08:42:44.797972 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:44 crc kubenswrapper[4873]: E1201 08:42:44.798360 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:45.298352486 +0000 UTC m=+141.200461025 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:44 crc kubenswrapper[4873]: W1201 08:42:44.800689 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod17e954cb_2d58_4616_bd5c_e813644d411d.slice/crio-a1f70a510294abd1431eca71ed8089791801ad1c58749a1b2f3a5d414c8533fd WatchSource:0}: Error finding container a1f70a510294abd1431eca71ed8089791801ad1c58749a1b2f3a5d414c8533fd: Status 404 returned error can't find the container with id a1f70a510294abd1431eca71ed8089791801ad1c58749a1b2f3a5d414c8533fd Dec 01 08:42:44 crc kubenswrapper[4873]: I1201 08:42:44.850235 4873 patch_prober.go:28] interesting pod/router-default-5444994796-h9kpk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 08:42:44 crc kubenswrapper[4873]: [-]has-synced failed: reason withheld Dec 01 08:42:44 crc kubenswrapper[4873]: [+]process-running ok Dec 01 08:42:44 crc kubenswrapper[4873]: healthz check failed Dec 01 08:42:44 crc kubenswrapper[4873]: I1201 08:42:44.850301 4873 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h9kpk" podUID="e686de65-3d86-4826-9f0a-4132c81e9ed2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 08:42:44 crc kubenswrapper[4873]: I1201 08:42:44.900031 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:44 crc kubenswrapper[4873]: E1201 08:42:44.900589 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:45.400562791 +0000 UTC m=+141.302671340 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:45 crc kubenswrapper[4873]: I1201 08:42:45.021579 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:45 crc kubenswrapper[4873]: E1201 08:42:45.022080 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:45.522066566 +0000 UTC m=+141.424175105 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:45 crc kubenswrapper[4873]: I1201 08:42:45.113320 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xfb2s"] Dec 01 08:42:45 crc kubenswrapper[4873]: I1201 08:42:45.134785 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:45 crc kubenswrapper[4873]: E1201 08:42:45.135239 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:45.635219784 +0000 UTC m=+141.537328323 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:45 crc kubenswrapper[4873]: I1201 08:42:45.144780 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-nd5kp"] Dec 01 08:42:45 crc kubenswrapper[4873]: I1201 08:42:45.172056 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-csmdk"] Dec 01 08:42:45 crc kubenswrapper[4873]: I1201 08:42:45.173685 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ptnr8" event={"ID":"f2208bec-45d1-45f0-8700-1218abfa7642","Type":"ContainerStarted","Data":"ac33ceda1978d3e0541ed6bea1a338af168a82c45dc5bc1495bd8f126d77e0b0"} Dec 01 08:42:45 crc kubenswrapper[4873]: I1201 08:42:45.194352 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7dcf8" event={"ID":"645ef95b-f94d-4ba4-8f10-fbe046a9d9a7","Type":"ContainerStarted","Data":"7a03686abd5905e26aa70a58867602bbb69572bfd6819cca91d641136f7a408f"} Dec 01 08:42:45 crc kubenswrapper[4873]: I1201 08:42:45.196919 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-292nl"] Dec 01 08:42:45 crc kubenswrapper[4873]: I1201 08:42:45.235976 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-w8xn2" event={"ID":"d0c705dd-9608-4b5e-8c91-2511d8427419","Type":"ContainerStarted","Data":"0b0c7ee62d72cc58d26f0eae185eb4dfe22d14a75379f76497d5aad1176315d4"} Dec 01 08:42:45 crc kubenswrapper[4873]: I1201 08:42:45.237659 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:45 crc kubenswrapper[4873]: I1201 08:42:45.240208 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409630-lnw2h"] Dec 01 08:42:45 crc kubenswrapper[4873]: E1201 08:42:45.248611 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:45.74856773 +0000 UTC m=+141.650676269 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:45 crc kubenswrapper[4873]: I1201 08:42:45.252113 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gscbx" event={"ID":"4fd09645-8bbe-458c-8d9b-b4c2875fd963","Type":"ContainerStarted","Data":"d42d66923aed493ff845216f2cc8fcb8995b5da86fad5b38fc06ab7fb40d0070"} Dec 01 08:42:45 crc kubenswrapper[4873]: I1201 08:42:45.254153 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-qvzhc" event={"ID":"17e954cb-2d58-4616-bd5c-e813644d411d","Type":"ContainerStarted","Data":"a1f70a510294abd1431eca71ed8089791801ad1c58749a1b2f3a5d414c8533fd"} Dec 01 08:42:45 crc kubenswrapper[4873]: I1201 08:42:45.258725 4873 patch_prober.go:28] interesting pod/downloads-7954f5f757-76drh container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Dec 01 08:42:45 crc kubenswrapper[4873]: I1201 08:42:45.258787 4873 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-76drh" podUID="dc0adee4-65bf-484b-8f55-3c83022cae82" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Dec 01 08:42:45 crc kubenswrapper[4873]: I1201 08:42:45.272344 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lcqht" Dec 01 08:42:45 crc kubenswrapper[4873]: I1201 08:42:45.278089 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cl5cd" Dec 01 08:42:45 crc kubenswrapper[4873]: I1201 08:42:45.337879 4873 patch_prober.go:28] interesting pod/router-default-5444994796-h9kpk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 08:42:45 crc kubenswrapper[4873]: [-]has-synced failed: reason withheld Dec 01 08:42:45 crc kubenswrapper[4873]: [+]process-running ok Dec 01 08:42:45 crc kubenswrapper[4873]: healthz check failed Dec 01 08:42:45 crc kubenswrapper[4873]: I1201 08:42:45.337961 4873 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h9kpk" podUID="e686de65-3d86-4826-9f0a-4132c81e9ed2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 08:42:45 crc kubenswrapper[4873]: I1201 08:42:45.339448 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:45 crc kubenswrapper[4873]: E1201 08:42:45.342486 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:45.842464921 +0000 UTC m=+141.744573460 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:45 crc kubenswrapper[4873]: I1201 08:42:45.396151 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xdpgx"] Dec 01 08:42:45 crc kubenswrapper[4873]: I1201 08:42:45.436846 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xrgx6"] Dec 01 08:42:45 crc kubenswrapper[4873]: I1201 08:42:45.441593 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:45 crc kubenswrapper[4873]: E1201 08:42:45.442103 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:45.942084289 +0000 UTC m=+141.844192818 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:45 crc kubenswrapper[4873]: I1201 08:42:45.442700 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-l9mq8"] Dec 01 08:42:45 crc kubenswrapper[4873]: I1201 08:42:45.470371 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-xt48p"] Dec 01 08:42:45 crc kubenswrapper[4873]: I1201 08:42:45.481695 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mvczl"] Dec 01 08:42:45 crc kubenswrapper[4873]: I1201 08:42:45.486932 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f5rc8"] Dec 01 08:42:45 crc kubenswrapper[4873]: I1201 08:42:45.502929 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-pwc6t"] Dec 01 08:42:45 crc kubenswrapper[4873]: I1201 08:42:45.505702 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-sg6bg"] Dec 01 08:42:45 crc kubenswrapper[4873]: I1201 08:42:45.521816 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q4wmp" Dec 01 08:42:45 crc kubenswrapper[4873]: I1201 08:42:45.534203 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-f5kts"] Dec 01 08:42:45 crc kubenswrapper[4873]: I1201 08:42:45.542317 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:45 crc kubenswrapper[4873]: E1201 08:42:45.542744 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:46.042724989 +0000 UTC m=+141.944833528 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:45 crc kubenswrapper[4873]: W1201 08:42:45.545381 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfb6d3b33_6120_40aa_8c4d_01f186d1ab17.slice/crio-2bf40f92de08367edfb7481c3e458b662bf2f972118fc801e84642ab0ee47534 WatchSource:0}: Error finding container 2bf40f92de08367edfb7481c3e458b662bf2f972118fc801e84642ab0ee47534: Status 404 returned error can't find the container with id 2bf40f92de08367edfb7481c3e458b662bf2f972118fc801e84642ab0ee47534 Dec 01 08:42:45 crc kubenswrapper[4873]: I1201 08:42:45.566116 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-gxhcq"] Dec 01 08:42:45 crc kubenswrapper[4873]: I1201 08:42:45.587898 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-s5ct4"] Dec 01 08:42:45 crc kubenswrapper[4873]: I1201 08:42:45.589801 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-9nmnq"] Dec 01 08:42:45 crc kubenswrapper[4873]: I1201 08:42:45.791326 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:45 crc kubenswrapper[4873]: E1201 08:42:45.801302 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:46.301082815 +0000 UTC m=+142.203191354 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:45 crc kubenswrapper[4873]: I1201 08:42:45.865722 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-5bhp6" Dec 01 08:42:45 crc kubenswrapper[4873]: I1201 08:42:45.867961 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-x9tsn"] Dec 01 08:42:45 crc kubenswrapper[4873]: I1201 08:42:45.898996 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:45 crc kubenswrapper[4873]: E1201 08:42:45.899984 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:46.399946591 +0000 UTC m=+142.302055130 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:45 crc kubenswrapper[4873]: I1201 08:42:45.901743 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:45 crc kubenswrapper[4873]: E1201 08:42:45.902201 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:46.402192114 +0000 UTC m=+142.304300653 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:45 crc kubenswrapper[4873]: I1201 08:42:45.917029 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7hstm"] Dec 01 08:42:45 crc kubenswrapper[4873]: I1201 08:42:45.934109 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-whl8s"] Dec 01 08:42:45 crc kubenswrapper[4873]: I1201 08:42:45.934199 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-crgj6"] Dec 01 08:42:45 crc kubenswrapper[4873]: I1201 08:42:45.947241 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-gvkmx"] Dec 01 08:42:45 crc kubenswrapper[4873]: I1201 08:42:45.953050 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-pp9bw"] Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.005709 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:46 crc kubenswrapper[4873]: E1201 08:42:46.005888 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:46.5058563 +0000 UTC m=+142.407964839 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.006371 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:46 crc kubenswrapper[4873]: E1201 08:42:46.006689 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:46.506674963 +0000 UTC m=+142.408783502 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.109155 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:46 crc kubenswrapper[4873]: E1201 08:42:46.109495 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:46.609445162 +0000 UTC m=+142.511553711 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.109755 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:46 crc kubenswrapper[4873]: E1201 08:42:46.110407 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:46.610396071 +0000 UTC m=+142.512504620 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:46 crc kubenswrapper[4873]: W1201 08:42:46.122403 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod12ddf317_c034_4154_9966_c8b502ec0d6c.slice/crio-f5685bb29a8cc493b6f9601ce622b6d05b5a8de08b5f712a07926baffaf2a4cb WatchSource:0}: Error finding container f5685bb29a8cc493b6f9601ce622b6d05b5a8de08b5f712a07926baffaf2a4cb: Status 404 returned error can't find the container with id f5685bb29a8cc493b6f9601ce622b6d05b5a8de08b5f712a07926baffaf2a4cb Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.212091 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:46 crc kubenswrapper[4873]: E1201 08:42:46.212430 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:46.712409358 +0000 UTC m=+142.614517897 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.291487 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-xt48p" event={"ID":"757ee90a-18e3-4ccb-b03c-43e90bf8cb34","Type":"ContainerStarted","Data":"5571762f140f55eb1e3ebf92d790a794d3250cebc28f19b8e0163f84a40aac18"} Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.315572 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:46 crc kubenswrapper[4873]: E1201 08:42:46.316267 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:46.816249201 +0000 UTC m=+142.718357740 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.322791 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-292nl" event={"ID":"9491cb1f-acf6-438b-8175-11737d7bd245","Type":"ContainerStarted","Data":"b078318cd9203fe5f24f2bd60ba9cea7d9e49405241d9f9008a2b5e2924098f8"} Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.322855 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-292nl" event={"ID":"9491cb1f-acf6-438b-8175-11737d7bd245","Type":"ContainerStarted","Data":"297abfad07bfd6ea781b5c866c64d514efa3d17ffa8ee136a685828d797de4a3"} Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.333644 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-f5kts" event={"ID":"67e07bc1-ce0f-453c-83ba-6f8cc36ed889","Type":"ContainerStarted","Data":"342342c3da7e1aa9d5f0b6997a02e40f454111d68f78669bf1286d9f4cf10cb9"} Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.370682 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xdpgx" event={"ID":"fb6d3b33-6120-40aa-8c4d-01f186d1ab17","Type":"ContainerStarted","Data":"2bf40f92de08367edfb7481c3e458b662bf2f972118fc801e84642ab0ee47534"} Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.378825 4873 patch_prober.go:28] interesting pod/router-default-5444994796-h9kpk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 08:42:46 crc kubenswrapper[4873]: [-]has-synced failed: reason withheld Dec 01 08:42:46 crc kubenswrapper[4873]: [+]process-running ok Dec 01 08:42:46 crc kubenswrapper[4873]: healthz check failed Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.378906 4873 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h9kpk" podUID="e686de65-3d86-4826-9f0a-4132c81e9ed2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.393641 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gscbx" event={"ID":"4fd09645-8bbe-458c-8d9b-b4c2875fd963","Type":"ContainerStarted","Data":"562d1e0f78b5dab3b11f60a810c061bbc34f7ef28087a07369c7fb19cee52c84"} Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.410625 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-292nl" podStartSLOduration=123.410607541 podStartE2EDuration="2m3.410607541s" podCreationTimestamp="2025-12-01 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:46.406134346 +0000 UTC m=+142.308242885" watchObservedRunningTime="2025-12-01 08:42:46.410607541 +0000 UTC m=+142.312716080" Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.412978 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-lnw2h" event={"ID":"83688ccb-ace0-4b0a-b584-e650671f5686","Type":"ContainerStarted","Data":"38b8fc7cc310a3ec72fe0ed1ba4dbfa51153edeb1fc2eb5f97423a83c30f7bc2"} Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.413061 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-lnw2h" event={"ID":"83688ccb-ace0-4b0a-b584-e650671f5686","Type":"ContainerStarted","Data":"c0555dcd08053d6bc41d0f8e5a5c15a6164f017939d68122c620aaa57bed825f"} Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.417298 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:46 crc kubenswrapper[4873]: E1201 08:42:46.418452 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:46.918433755 +0000 UTC m=+142.820542294 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.437359 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gscbx" podStartSLOduration=123.437336899 podStartE2EDuration="2m3.437336899s" podCreationTimestamp="2025-12-01 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:46.427065193 +0000 UTC m=+142.329173752" watchObservedRunningTime="2025-12-01 08:42:46.437336899 +0000 UTC m=+142.339445438" Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.452310 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7dcf8" event={"ID":"645ef95b-f94d-4ba4-8f10-fbe046a9d9a7","Type":"ContainerStarted","Data":"8d8fe6be03633862462cea6c27da55cd21237aa926023cf05417b6d50f43112b"} Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.454683 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-lnw2h" podStartSLOduration=123.454649966 podStartE2EDuration="2m3.454649966s" podCreationTimestamp="2025-12-01 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:46.450464462 +0000 UTC m=+142.352573001" watchObservedRunningTime="2025-12-01 08:42:46.454649966 +0000 UTC m=+142.356758505" Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.472453 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9nmnq" event={"ID":"04b9f090-d8f6-4122-bbff-af681fcc4f42","Type":"ContainerStarted","Data":"630ae035afbaf4b9f91af013ad2f5b29e49eab5084fd49d10db2158c5aa823f7"} Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.474255 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f5rc8" event={"ID":"90dd16fd-d069-4007-8c82-4811cb343380","Type":"ContainerStarted","Data":"7cceefc5b28f1ba897aa9fbc845cbe89e7e5f330fe8d1ba9af751bfe1584b31a"} Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.475329 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-x9tsn" event={"ID":"812f722d-c159-40bb-8901-bac36d182280","Type":"ContainerStarted","Data":"50371f573fc71d6c6ae6cb9ec6fed0291eb86ba69fc50a88cebcff1607d028b0"} Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.479432 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xfb2s" event={"ID":"ae767d9a-3df4-413a-bc25-8fc819c7088e","Type":"ContainerStarted","Data":"b7a1c359058c1a188ebfcc39c97ea810d57f54bb84c871edf5c2ccf6cb49cab7"} Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.479469 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xfb2s" event={"ID":"ae767d9a-3df4-413a-bc25-8fc819c7088e","Type":"ContainerStarted","Data":"ef73e603c26413e7b209800cabd4c8a5a210542fd26ac8de65969dd29f4d1ba6"} Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.480506 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-pp9bw" event={"ID":"12ddf317-c034-4154-9966-c8b502ec0d6c","Type":"ContainerStarted","Data":"f5685bb29a8cc493b6f9601ce622b6d05b5a8de08b5f712a07926baffaf2a4cb"} Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.491970 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-qvzhc" event={"ID":"17e954cb-2d58-4616-bd5c-e813644d411d","Type":"ContainerStarted","Data":"f805d466a8e553b0a0dcca4074e83f2b28dfd46086427c2c2809fab6425393f8"} Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.504223 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7dcf8" podStartSLOduration=123.504195549 podStartE2EDuration="2m3.504195549s" podCreationTimestamp="2025-12-01 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:46.501772479 +0000 UTC m=+142.403881028" watchObservedRunningTime="2025-12-01 08:42:46.504195549 +0000 UTC m=+142.406304088" Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.518505 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:46 crc kubenswrapper[4873]: E1201 08:42:46.519924 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:47.01990588 +0000 UTC m=+142.922014419 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.555934 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-gvkmx" event={"ID":"0726d810-1988-4eb3-9077-e8892670dc8c","Type":"ContainerStarted","Data":"24a9b2be4599ad585f2549cf6bf40d9114a764b0362f1ecc4c9563edc22c6b48"} Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.564844 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-csmdk" event={"ID":"f99594fb-adb1-469b-960f-2a56dc3427bd","Type":"ContainerStarted","Data":"428be3ecf9f91e57d625bdbbb5cae6683d1a4b4f3a1dd2329d4dd7df989a706a"} Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.564988 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-csmdk" event={"ID":"f99594fb-adb1-469b-960f-2a56dc3427bd","Type":"ContainerStarted","Data":"a4321da775982296ca0c05e827de2520170b5b68d7bb715dd4b809273bbf65f4"} Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.576516 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mvczl" event={"ID":"2a86c401-c7dd-4b95-b267-8c58e5728e6a","Type":"ContainerStarted","Data":"53f1ccdc5975a3044bfaa98b265a3160b487fe556bdaaa1d66ce3ff460f91e73"} Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.579002 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xfb2s" podStartSLOduration=123.578984778 podStartE2EDuration="2m3.578984778s" podCreationTimestamp="2025-12-01 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:46.576989275 +0000 UTC m=+142.479097814" watchObservedRunningTime="2025-12-01 08:42:46.578984778 +0000 UTC m=+142.481093307" Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.621747 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-csmdk" podStartSLOduration=123.621717609 podStartE2EDuration="2m3.621717609s" podCreationTimestamp="2025-12-01 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:46.61571299 +0000 UTC m=+142.517821529" watchObservedRunningTime="2025-12-01 08:42:46.621717609 +0000 UTC m=+142.523826158" Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.624196 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:46 crc kubenswrapper[4873]: E1201 08:42:46.624764 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:47.124734113 +0000 UTC m=+143.026842652 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.624858 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.625169 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ptnr8" event={"ID":"f2208bec-45d1-45f0-8700-1218abfa7642","Type":"ContainerStarted","Data":"41e607751919a1d20eb15100959c9081d8dbc99f667398de6d2e9326228931be"} Dec 01 08:42:46 crc kubenswrapper[4873]: E1201 08:42:46.625223 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:47.125211773 +0000 UTC m=+143.027320312 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.638224 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-whl8s" event={"ID":"3dc1fc6b-9c79-4a78-af19-d49195bc94be","Type":"ContainerStarted","Data":"c724691353505e065cb189d506ac82c14fd4e7a79be4550785c3856f207aaea2"} Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.705463 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xrgx6" event={"ID":"e3e97e3b-f082-4093-b183-fcd5ac857c07","Type":"ContainerStarted","Data":"1363d975777f84ccdf9ad3426a40223567eb235680f8761de3a22a5d0baf29a7"} Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.705950 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xrgx6" Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.721249 4873 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-xrgx6 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" start-of-body= Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.721371 4873 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xrgx6" podUID="e3e97e3b-f082-4093-b183-fcd5ac857c07" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.723641 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7hstm" event={"ID":"2ea695f7-4e9e-4b12-8e0c-22431291576d","Type":"ContainerStarted","Data":"4813d2fb668949029e06c337a682b5bbc38ebdf66e908a6ff62a80f644365233"} Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.729141 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.737280 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-pwc6t" event={"ID":"23e5ea75-df3a-4a09-87f4-542f0063f4aa","Type":"ContainerStarted","Data":"6abaf6693fa16ad8de2593384903cf3c31805c63128d877f58d36e16511d090b"} Dec 01 08:42:46 crc kubenswrapper[4873]: E1201 08:42:46.737412 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:47.237371931 +0000 UTC m=+143.139480460 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.738934 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:46 crc kubenswrapper[4873]: E1201 08:42:46.741886 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:47.241802865 +0000 UTC m=+143.143911404 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.748746 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ptnr8" podStartSLOduration=123.748724151 podStartE2EDuration="2m3.748724151s" podCreationTimestamp="2025-12-01 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:46.671087684 +0000 UTC m=+142.573196223" watchObservedRunningTime="2025-12-01 08:42:46.748724151 +0000 UTC m=+142.650832680" Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.787792 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-pwc6t" podStartSLOduration=8.787734388 podStartE2EDuration="8.787734388s" podCreationTimestamp="2025-12-01 08:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:46.782094694 +0000 UTC m=+142.684203233" watchObservedRunningTime="2025-12-01 08:42:46.787734388 +0000 UTC m=+142.689842927" Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.790740 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xrgx6" podStartSLOduration=123.790659139 podStartE2EDuration="2m3.790659139s" podCreationTimestamp="2025-12-01 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:46.762665909 +0000 UTC m=+142.664774448" watchObservedRunningTime="2025-12-01 08:42:46.790659139 +0000 UTC m=+142.692767678" Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.844361 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-crgj6" event={"ID":"9a6d7d20-0eea-42ae-92e0-91b1b555bf6f","Type":"ContainerStarted","Data":"c2334b1a8a516a9578f9dee1784e122f90f4920f91caf3bde9181121ca247d6e"} Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.845361 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:46 crc kubenswrapper[4873]: E1201 08:42:46.845464 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:47.345442799 +0000 UTC m=+143.247551338 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.846132 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:46 crc kubenswrapper[4873]: E1201 08:42:46.846535 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:47.346519044 +0000 UTC m=+143.248627583 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.875346 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s5ct4" event={"ID":"7aee65fe-0d01-4db7-95ee-131e6ead9071","Type":"ContainerStarted","Data":"718e95edcf828271055e192f718a1a024c1a78d2df4e16e264ef56e778d7dbb2"} Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.879898 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-l9mq8" event={"ID":"dfbcf0e0-e5a1-41e0-813c-008f355a1f0a","Type":"ContainerStarted","Data":"daa01de2ef543b92917c70007965469744431e74771c1053965e03f948f49890"} Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.885372 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" event={"ID":"2404e1f8-203a-4055-8f66-21d8867e7a92","Type":"ContainerStarted","Data":"1ba6bf355d3416ca5e8e0786258527e32267090fd584943d2badfd06124aff1f"} Dec 01 08:42:46 crc kubenswrapper[4873]: I1201 08:42:46.955876 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:46 crc kubenswrapper[4873]: E1201 08:42:46.956299 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:47.456275712 +0000 UTC m=+143.358384251 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:47 crc kubenswrapper[4873]: I1201 08:42:47.063158 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:47 crc kubenswrapper[4873]: E1201 08:42:47.063575 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:47.563560797 +0000 UTC m=+143.465669336 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:47 crc kubenswrapper[4873]: I1201 08:42:47.139748 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-w8xn2" event={"ID":"d0c705dd-9608-4b5e-8c91-2511d8427419","Type":"ContainerStarted","Data":"f63ffef129ab5e39cd2ad381167ab852d23350d01963ee13c44f1856c3033ee9"} Dec 01 08:42:47 crc kubenswrapper[4873]: I1201 08:42:47.157503 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gxhcq" event={"ID":"4af31a76-3090-497d-b39f-965c27a27a1c","Type":"ContainerStarted","Data":"f32eb0436cb58cd4101a1d14bfcfe6fa84bd22e74ca20eb335dfdb648452a327"} Dec 01 08:42:47 crc kubenswrapper[4873]: I1201 08:42:47.167058 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:47 crc kubenswrapper[4873]: E1201 08:42:47.168084 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:47.668064497 +0000 UTC m=+143.570173036 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:47 crc kubenswrapper[4873]: I1201 08:42:47.176628 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-nd5kp" event={"ID":"10d7dabb-aa6f-4f2b-acf9-b02dc369d8be","Type":"ContainerStarted","Data":"b1613e2aa4e6ca4bad7e7799741598426fb4351426563fe071ad7b525eb9764b"} Dec 01 08:42:47 crc kubenswrapper[4873]: I1201 08:42:47.176702 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-nd5kp" event={"ID":"10d7dabb-aa6f-4f2b-acf9-b02dc369d8be","Type":"ContainerStarted","Data":"12b9a4e7cd3f051bb25ab96f40714567d53cbe027267cb04347d3d3f6eb4d3db"} Dec 01 08:42:47 crc kubenswrapper[4873]: I1201 08:42:47.269427 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:47 crc kubenswrapper[4873]: E1201 08:42:47.273200 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:47.773180783 +0000 UTC m=+143.675289392 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:47 crc kubenswrapper[4873]: I1201 08:42:47.340415 4873 patch_prober.go:28] interesting pod/router-default-5444994796-h9kpk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 08:42:47 crc kubenswrapper[4873]: [-]has-synced failed: reason withheld Dec 01 08:42:47 crc kubenswrapper[4873]: [+]process-running ok Dec 01 08:42:47 crc kubenswrapper[4873]: healthz check failed Dec 01 08:42:47 crc kubenswrapper[4873]: I1201 08:42:47.340486 4873 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h9kpk" podUID="e686de65-3d86-4826-9f0a-4132c81e9ed2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 08:42:47 crc kubenswrapper[4873]: I1201 08:42:47.372190 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:47 crc kubenswrapper[4873]: E1201 08:42:47.373032 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:47.872995779 +0000 UTC m=+143.775104318 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:47 crc kubenswrapper[4873]: I1201 08:42:47.473230 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:47 crc kubenswrapper[4873]: E1201 08:42:47.473837 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:47.973816027 +0000 UTC m=+143.875924566 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:47 crc kubenswrapper[4873]: I1201 08:42:47.590291 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:47 crc kubenswrapper[4873]: E1201 08:42:47.591503 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:48.091467412 +0000 UTC m=+143.993576121 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:47 crc kubenswrapper[4873]: I1201 08:42:47.693628 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:47 crc kubenswrapper[4873]: E1201 08:42:47.694030 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:48.19399987 +0000 UTC m=+144.096108409 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:47 crc kubenswrapper[4873]: I1201 08:42:47.794450 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:47 crc kubenswrapper[4873]: E1201 08:42:47.795350 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:48.295328469 +0000 UTC m=+144.197437008 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:47 crc kubenswrapper[4873]: E1201 08:42:47.899055 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:48.399037886 +0000 UTC m=+144.301146425 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:47 crc kubenswrapper[4873]: I1201 08:42:47.898603 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.018633 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:48 crc kubenswrapper[4873]: E1201 08:42:48.019695 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:48.519670255 +0000 UTC m=+144.421778794 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.120581 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:48 crc kubenswrapper[4873]: E1201 08:42:48.121482 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:48.621468713 +0000 UTC m=+144.523577252 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.223131 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:48 crc kubenswrapper[4873]: E1201 08:42:48.223521 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:48.72347913 +0000 UTC m=+144.625587679 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.241807 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xdpgx" event={"ID":"fb6d3b33-6120-40aa-8c4d-01f186d1ab17","Type":"ContainerStarted","Data":"9e2ed40ad6c5c2dadb8c3f491f887442af6087e03e90f717df782ed16209c273"} Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.253936 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s5ct4" event={"ID":"7aee65fe-0d01-4db7-95ee-131e6ead9071","Type":"ContainerStarted","Data":"335584c1e7cf9631b9b2dcd7dc7e9dc1e960db3c62ffe67f9e1f057cb34ea208"} Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.253986 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s5ct4" event={"ID":"7aee65fe-0d01-4db7-95ee-131e6ead9071","Type":"ContainerStarted","Data":"ef8a00dcaf6d6a7018299fa61bcbdd0c1f0fecd1f5f1196990fc0334bf053b8f"} Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.264066 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-w8xn2" podStartSLOduration=125.264035361 podStartE2EDuration="2m5.264035361s" podCreationTimestamp="2025-12-01 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:47.292637559 +0000 UTC m=+143.194746098" watchObservedRunningTime="2025-12-01 08:42:48.264035361 +0000 UTC m=+144.166143910" Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.267810 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mvczl" event={"ID":"2a86c401-c7dd-4b95-b267-8c58e5728e6a","Type":"ContainerStarted","Data":"0d0d8b1119477758b4ead49e868fccf2f70723626ee464fb87f26649d1f4f57a"} Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.267860 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mvczl" event={"ID":"2a86c401-c7dd-4b95-b267-8c58e5728e6a","Type":"ContainerStarted","Data":"33a3e9b4bacc7d330d3b3ec42773a78d334e53d9e4e781e4b187115e2621bde6"} Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.268669 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mvczl" Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.271474 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gxhcq" event={"ID":"4af31a76-3090-497d-b39f-965c27a27a1c","Type":"ContainerStarted","Data":"0a34f55de871041fe295fd6f024acfc4a22d59b0526e7b83d947008fa23a3de1"} Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.271506 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gxhcq" event={"ID":"4af31a76-3090-497d-b39f-965c27a27a1c","Type":"ContainerStarted","Data":"456a51ca7ed33adb04d853c2167ca7033959386245af1103a31a147f27c78a1b"} Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.273223 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-l9mq8" event={"ID":"dfbcf0e0-e5a1-41e0-813c-008f355a1f0a","Type":"ContainerStarted","Data":"a3aa95e865401800bd63cdbe3ed77a93e395c9fec3c98ad7cda1ab077e106b78"} Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.274268 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7hstm" event={"ID":"2ea695f7-4e9e-4b12-8e0c-22431291576d","Type":"ContainerStarted","Data":"ac126d7386d526b4abbff206ecba9a518fe214fc8803ff9777184f9779560f9d"} Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.275136 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-7hstm" Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.276738 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xrgx6" event={"ID":"e3e97e3b-f082-4093-b183-fcd5ac857c07","Type":"ContainerStarted","Data":"bf51e3d7a260620ca29fa33cc2197256300ecc5951234404b619bb8436172d58"} Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.283789 4873 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-xrgx6 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" start-of-body= Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.283849 4873 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xrgx6" podUID="e3e97e3b-f082-4093-b183-fcd5ac857c07" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.288157 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-x9tsn" event={"ID":"812f722d-c159-40bb-8901-bac36d182280","Type":"ContainerStarted","Data":"9aff4329804aab9898adebdb8d8c5beba510ceb1fef93309613c516ca574fe18"} Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.290288 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-pp9bw" event={"ID":"12ddf317-c034-4154-9966-c8b502ec0d6c","Type":"ContainerStarted","Data":"1548de3b2f6f45667ebc07199460d0a0dbabb16ce6533cf62ae5b1f0c84e52c7"} Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.299696 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xdpgx" podStartSLOduration=125.299673037 podStartE2EDuration="2m5.299673037s" podCreationTimestamp="2025-12-01 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:48.267426021 +0000 UTC m=+144.169534560" watchObservedRunningTime="2025-12-01 08:42:48.299673037 +0000 UTC m=+144.201781576" Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.300835 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s5ct4" podStartSLOduration=125.300829295 podStartE2EDuration="2m5.300829295s" podCreationTimestamp="2025-12-01 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:48.297983227 +0000 UTC m=+144.200091766" watchObservedRunningTime="2025-12-01 08:42:48.300829295 +0000 UTC m=+144.202937834" Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.301534 4873 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-7hstm container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.301572 4873 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-7hstm" podUID="2ea695f7-4e9e-4b12-8e0c-22431291576d" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.305803 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f5rc8" event={"ID":"90dd16fd-d069-4007-8c82-4811cb343380","Type":"ContainerStarted","Data":"ba7ffa1b4369ef7b9882d01145dd501a6371ec8c45fe8b0b5dbbf09307a0af89"} Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.306851 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f5rc8" Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.309787 4873 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-f5rc8 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:5443/healthz\": dial tcp 10.217.0.32:5443: connect: connection refused" start-of-body= Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.309823 4873 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f5rc8" podUID="90dd16fd-d069-4007-8c82-4811cb343380" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.32:5443/healthz\": dial tcp 10.217.0.32:5443: connect: connection refused" Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.315309 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-nd5kp" event={"ID":"10d7dabb-aa6f-4f2b-acf9-b02dc369d8be","Type":"ContainerStarted","Data":"c2cdc997da12f8bca7c5372fd7ea9c0d950d80f9ee29f35620f2f8ff26437407"} Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.318063 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-qvzhc" event={"ID":"17e954cb-2d58-4616-bd5c-e813644d411d","Type":"ContainerStarted","Data":"72d7d9123d514b7854b26ee7159b7d8fc329b5f5f8aa1f9224715a28fcd52b8a"} Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.320297 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-gvkmx" event={"ID":"0726d810-1988-4eb3-9077-e8892670dc8c","Type":"ContainerStarted","Data":"6c8921ba096335a6cb6f27fed38c23b246745ee4aaf1da7c96b907d7bc509cc6"} Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.322296 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-xt48p" event={"ID":"757ee90a-18e3-4ccb-b03c-43e90bf8cb34","Type":"ContainerStarted","Data":"a82857671955569dc61db2301570d26e350b8ae6923e25ccd18ee9d2a5e3a04b"} Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.325325 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:48 crc kubenswrapper[4873]: E1201 08:42:48.329666 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:48.829644629 +0000 UTC m=+144.731753378 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.329897 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-crgj6" event={"ID":"9a6d7d20-0eea-42ae-92e0-91b1b555bf6f","Type":"ContainerStarted","Data":"06bf2af2b6371fffcfe9b40add9d7fa6453eff3f100a7980520cd3ac8df78d06"} Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.338217 4873 patch_prober.go:28] interesting pod/router-default-5444994796-h9kpk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 08:42:48 crc kubenswrapper[4873]: [-]has-synced failed: reason withheld Dec 01 08:42:48 crc kubenswrapper[4873]: [+]process-running ok Dec 01 08:42:48 crc kubenswrapper[4873]: healthz check failed Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.338282 4873 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h9kpk" podUID="e686de65-3d86-4826-9f0a-4132c81e9ed2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.350694 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" event={"ID":"2404e1f8-203a-4055-8f66-21d8867e7a92","Type":"ContainerStarted","Data":"a3eb595bf52712b15c741f24833ada335abb9c444fd6b09b00e8a9b21a9e3042"} Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.352167 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.354407 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-7hstm" podStartSLOduration=125.354383704 podStartE2EDuration="2m5.354383704s" podCreationTimestamp="2025-12-01 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:48.350642839 +0000 UTC m=+144.252751378" watchObservedRunningTime="2025-12-01 08:42:48.354383704 +0000 UTC m=+144.256492243" Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.361875 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-whl8s" event={"ID":"3dc1fc6b-9c79-4a78-af19-d49195bc94be","Type":"ContainerStarted","Data":"bfcdba2b2687812af621a1be360d277231812a2e9d765cae071540b023f62bd4"} Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.362901 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-whl8s" Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.363439 4873 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-sg6bg container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.21:6443/healthz\": dial tcp 10.217.0.21:6443: connect: connection refused" start-of-body= Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.363500 4873 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" podUID="2404e1f8-203a-4055-8f66-21d8867e7a92" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.21:6443/healthz\": dial tcp 10.217.0.21:6443: connect: connection refused" Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.364545 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-f5kts" event={"ID":"67e07bc1-ce0f-453c-83ba-6f8cc36ed889","Type":"ContainerStarted","Data":"53a6d555617a4e8917f511b5a06f54588d320b571384eb542c39960972105d62"} Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.369422 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-pwc6t" event={"ID":"23e5ea75-df3a-4a09-87f4-542f0063f4aa","Type":"ContainerStarted","Data":"c32ad1bc5983dc37f11928d22708ff673af2046b7bcda32e2df1b80ddb7c65d2"} Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.388162 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gxhcq" podStartSLOduration=125.388126723 podStartE2EDuration="2m5.388126723s" podCreationTimestamp="2025-12-01 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:48.375438917 +0000 UTC m=+144.277547456" watchObservedRunningTime="2025-12-01 08:42:48.388126723 +0000 UTC m=+144.290235262" Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.402303 4873 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-whl8s container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" start-of-body= Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.402381 4873 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-whl8s" podUID="3dc1fc6b-9c79-4a78-af19-d49195bc94be" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.426934 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:48 crc kubenswrapper[4873]: E1201 08:42:48.427110 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:48.927086307 +0000 UTC m=+144.829194846 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.433725 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:48 crc kubenswrapper[4873]: E1201 08:42:48.516427 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:49.016392447 +0000 UTC m=+144.918500986 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.533625 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-x9tsn" podStartSLOduration=125.533607171 podStartE2EDuration="2m5.533607171s" podCreationTimestamp="2025-12-01 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:48.432262741 +0000 UTC m=+144.334371270" watchObservedRunningTime="2025-12-01 08:42:48.533607171 +0000 UTC m=+144.435715710" Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.534536 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mvczl" podStartSLOduration=125.534533059 podStartE2EDuration="2m5.534533059s" podCreationTimestamp="2025-12-01 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:48.532189262 +0000 UTC m=+144.434297801" watchObservedRunningTime="2025-12-01 08:42:48.534533059 +0000 UTC m=+144.436641598" Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.535335 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:48 crc kubenswrapper[4873]: E1201 08:42:48.535614 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:49.035603643 +0000 UTC m=+144.937712182 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.637569 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:48 crc kubenswrapper[4873]: E1201 08:42:48.638475 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:49.138459015 +0000 UTC m=+145.040567554 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.766180 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:48 crc kubenswrapper[4873]: E1201 08:42:48.766585 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:49.266565683 +0000 UTC m=+145.168674222 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.801648 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-pp9bw" podStartSLOduration=125.801626206 podStartE2EDuration="2m5.801626206s" podCreationTimestamp="2025-12-01 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:48.766999741 +0000 UTC m=+144.669108280" watchObservedRunningTime="2025-12-01 08:42:48.801626206 +0000 UTC m=+144.703734745" Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.802256 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-xt48p" podStartSLOduration=125.802250571 podStartE2EDuration="2m5.802250571s" podCreationTimestamp="2025-12-01 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:48.800147094 +0000 UTC m=+144.702255633" watchObservedRunningTime="2025-12-01 08:42:48.802250571 +0000 UTC m=+144.704359110" Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.842507 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-nd5kp" podStartSLOduration=125.842486699 podStartE2EDuration="2m5.842486699s" podCreationTimestamp="2025-12-01 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:48.842182126 +0000 UTC m=+144.744290665" watchObservedRunningTime="2025-12-01 08:42:48.842486699 +0000 UTC m=+144.744595238" Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.870292 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:48 crc kubenswrapper[4873]: E1201 08:42:48.870728 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:49.370714888 +0000 UTC m=+145.272823427 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.900953 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-qvzhc" podStartSLOduration=125.900934901 podStartE2EDuration="2m5.900934901s" podCreationTimestamp="2025-12-01 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:48.863258179 +0000 UTC m=+144.765366718" watchObservedRunningTime="2025-12-01 08:42:48.900934901 +0000 UTC m=+144.803043440" Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.962917 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" podStartSLOduration=125.962892018 podStartE2EDuration="2m5.962892018s" podCreationTimestamp="2025-12-01 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:48.904917706 +0000 UTC m=+144.807026245" watchObservedRunningTime="2025-12-01 08:42:48.962892018 +0000 UTC m=+144.865000557" Dec 01 08:42:48 crc kubenswrapper[4873]: I1201 08:42:48.979008 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:48 crc kubenswrapper[4873]: E1201 08:42:48.979410 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:49.479390891 +0000 UTC m=+145.381499430 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:49 crc kubenswrapper[4873]: I1201 08:42:49.007427 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-whl8s" podStartSLOduration=126.007404912 podStartE2EDuration="2m6.007404912s" podCreationTimestamp="2025-12-01 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:48.963236072 +0000 UTC m=+144.865344611" watchObservedRunningTime="2025-12-01 08:42:49.007404912 +0000 UTC m=+144.909513451" Dec 01 08:42:49 crc kubenswrapper[4873]: I1201 08:42:49.008186 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f5rc8" podStartSLOduration=126.008182244 podStartE2EDuration="2m6.008182244s" podCreationTimestamp="2025-12-01 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:49.005252793 +0000 UTC m=+144.907361332" watchObservedRunningTime="2025-12-01 08:42:49.008182244 +0000 UTC m=+144.910290783" Dec 01 08:42:49 crc kubenswrapper[4873]: I1201 08:42:49.080711 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:49 crc kubenswrapper[4873]: E1201 08:42:49.081128 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:49.581113047 +0000 UTC m=+145.483221586 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:49 crc kubenswrapper[4873]: I1201 08:42:49.183838 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:49 crc kubenswrapper[4873]: E1201 08:42:49.184332 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:49.684293032 +0000 UTC m=+145.586401581 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:49 crc kubenswrapper[4873]: I1201 08:42:49.184660 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:49 crc kubenswrapper[4873]: E1201 08:42:49.185004 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:49.684991651 +0000 UTC m=+145.587100190 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:49 crc kubenswrapper[4873]: I1201 08:42:49.286762 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:49 crc kubenswrapper[4873]: E1201 08:42:49.287191 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:49.787168085 +0000 UTC m=+145.689276634 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:49 crc kubenswrapper[4873]: I1201 08:42:49.287280 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:49 crc kubenswrapper[4873]: E1201 08:42:49.287747 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:49.787729998 +0000 UTC m=+145.689838537 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:49 crc kubenswrapper[4873]: I1201 08:42:49.347527 4873 patch_prober.go:28] interesting pod/router-default-5444994796-h9kpk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 08:42:49 crc kubenswrapper[4873]: [-]has-synced failed: reason withheld Dec 01 08:42:49 crc kubenswrapper[4873]: [+]process-running ok Dec 01 08:42:49 crc kubenswrapper[4873]: healthz check failed Dec 01 08:42:49 crc kubenswrapper[4873]: I1201 08:42:49.347626 4873 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h9kpk" podUID="e686de65-3d86-4826-9f0a-4132c81e9ed2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 08:42:49 crc kubenswrapper[4873]: I1201 08:42:49.394962 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:49 crc kubenswrapper[4873]: E1201 08:42:49.395452 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:49.895431551 +0000 UTC m=+145.797540080 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:49 crc kubenswrapper[4873]: I1201 08:42:49.433179 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-f5kts" event={"ID":"67e07bc1-ce0f-453c-83ba-6f8cc36ed889","Type":"ContainerStarted","Data":"d7c59bde9cb300c5f86d516948e65401442652daf7a37ff791acf809faa6c541"} Dec 01 08:42:49 crc kubenswrapper[4873]: I1201 08:42:49.438046 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-crgj6" event={"ID":"9a6d7d20-0eea-42ae-92e0-91b1b555bf6f","Type":"ContainerStarted","Data":"c0918e743d7f038cb07b496ece1e9c31472465175a448e76ea321bf5e06aa16f"} Dec 01 08:42:49 crc kubenswrapper[4873]: I1201 08:42:49.438314 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-crgj6" Dec 01 08:42:49 crc kubenswrapper[4873]: I1201 08:42:49.439324 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9nmnq" event={"ID":"04b9f090-d8f6-4122-bbff-af681fcc4f42","Type":"ContainerStarted","Data":"6c8e90f024a4bc580e2c5bb8b60513d5177e8e315794b053e45e7e0b0849dddd"} Dec 01 08:42:49 crc kubenswrapper[4873]: I1201 08:42:49.440692 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-gvkmx" event={"ID":"0726d810-1988-4eb3-9077-e8892670dc8c","Type":"ContainerStarted","Data":"bd40d4f752051d79cfae8d0cfde78b50278a0de83d90b49390cbdb2770cb5500"} Dec 01 08:42:49 crc kubenswrapper[4873]: I1201 08:42:49.443951 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-l9mq8" event={"ID":"dfbcf0e0-e5a1-41e0-813c-008f355a1f0a","Type":"ContainerStarted","Data":"2902583f5fa92574cfd561c216b30258864ae3c3d2bfc4b3cedf5000c9395d25"} Dec 01 08:42:49 crc kubenswrapper[4873]: I1201 08:42:49.445077 4873 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-whl8s container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" start-of-body= Dec 01 08:42:49 crc kubenswrapper[4873]: I1201 08:42:49.445119 4873 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-whl8s" podUID="3dc1fc6b-9c79-4a78-af19-d49195bc94be" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" Dec 01 08:42:49 crc kubenswrapper[4873]: I1201 08:42:49.445832 4873 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-7hstm container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Dec 01 08:42:49 crc kubenswrapper[4873]: I1201 08:42:49.445941 4873 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-7hstm" podUID="2ea695f7-4e9e-4b12-8e0c-22431291576d" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" Dec 01 08:42:49 crc kubenswrapper[4873]: I1201 08:42:49.446998 4873 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-sg6bg container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.21:6443/healthz\": dial tcp 10.217.0.21:6443: connect: connection refused" start-of-body= Dec 01 08:42:49 crc kubenswrapper[4873]: I1201 08:42:49.447042 4873 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" podUID="2404e1f8-203a-4055-8f66-21d8867e7a92" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.21:6443/healthz\": dial tcp 10.217.0.21:6443: connect: connection refused" Dec 01 08:42:49 crc kubenswrapper[4873]: I1201 08:42:49.497928 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:49 crc kubenswrapper[4873]: E1201 08:42:49.501926 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:50.001905993 +0000 UTC m=+145.904014532 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:49 crc kubenswrapper[4873]: I1201 08:42:49.530087 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-l9mq8" podStartSLOduration=126.53006612 podStartE2EDuration="2m6.53006612s" podCreationTimestamp="2025-12-01 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:49.52983455 +0000 UTC m=+145.431943079" watchObservedRunningTime="2025-12-01 08:42:49.53006612 +0000 UTC m=+145.432174659" Dec 01 08:42:49 crc kubenswrapper[4873]: I1201 08:42:49.531865 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-f5kts" podStartSLOduration=126.531859814 podStartE2EDuration="2m6.531859814s" podCreationTimestamp="2025-12-01 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:49.469147835 +0000 UTC m=+145.371256375" watchObservedRunningTime="2025-12-01 08:42:49.531859814 +0000 UTC m=+145.433968353" Dec 01 08:42:49 crc kubenswrapper[4873]: I1201 08:42:49.557533 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-gvkmx" podStartSLOduration=126.557515167 podStartE2EDuration="2m6.557515167s" podCreationTimestamp="2025-12-01 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:49.555520394 +0000 UTC m=+145.457628943" watchObservedRunningTime="2025-12-01 08:42:49.557515167 +0000 UTC m=+145.459623706" Dec 01 08:42:49 crc kubenswrapper[4873]: I1201 08:42:49.608461 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:49 crc kubenswrapper[4873]: E1201 08:42:49.608880 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:50.108860915 +0000 UTC m=+146.010969454 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:49 crc kubenswrapper[4873]: I1201 08:42:49.648502 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xrgx6" Dec 01 08:42:49 crc kubenswrapper[4873]: I1201 08:42:49.686492 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-crgj6" podStartSLOduration=11.68646799 podStartE2EDuration="11.68646799s" podCreationTimestamp="2025-12-01 08:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:49.58800414 +0000 UTC m=+145.490112679" watchObservedRunningTime="2025-12-01 08:42:49.68646799 +0000 UTC m=+145.588576529" Dec 01 08:42:49 crc kubenswrapper[4873]: I1201 08:42:49.711063 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:49 crc kubenswrapper[4873]: E1201 08:42:49.711738 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:50.211722447 +0000 UTC m=+146.113830986 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:49 crc kubenswrapper[4873]: I1201 08:42:49.814408 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:49 crc kubenswrapper[4873]: E1201 08:42:49.814732 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:50.314712114 +0000 UTC m=+146.216820653 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:49 crc kubenswrapper[4873]: I1201 08:42:49.943578 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:49 crc kubenswrapper[4873]: E1201 08:42:49.944035 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:50.444005462 +0000 UTC m=+146.346114001 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:50 crc kubenswrapper[4873]: I1201 08:42:50.044885 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:50 crc kubenswrapper[4873]: E1201 08:42:50.045394 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:50.545366242 +0000 UTC m=+146.447474781 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:50 crc kubenswrapper[4873]: I1201 08:42:50.055541 4873 patch_prober.go:28] interesting pod/downloads-7954f5f757-76drh container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Dec 01 08:42:50 crc kubenswrapper[4873]: I1201 08:42:50.055631 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-76drh" podUID="dc0adee4-65bf-484b-8f55-3c83022cae82" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Dec 01 08:42:50 crc kubenswrapper[4873]: I1201 08:42:50.056043 4873 patch_prober.go:28] interesting pod/downloads-7954f5f757-76drh container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Dec 01 08:42:50 crc kubenswrapper[4873]: I1201 08:42:50.056065 4873 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-76drh" podUID="dc0adee4-65bf-484b-8f55-3c83022cae82" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Dec 01 08:42:50 crc kubenswrapper[4873]: I1201 08:42:50.153627 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:50 crc kubenswrapper[4873]: E1201 08:42:50.154109 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:50.654091687 +0000 UTC m=+146.556200226 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:50 crc kubenswrapper[4873]: I1201 08:42:50.333229 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:50 crc kubenswrapper[4873]: E1201 08:42:50.333717 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:50.833693449 +0000 UTC m=+146.735801998 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:50 crc kubenswrapper[4873]: I1201 08:42:50.336207 4873 patch_prober.go:28] interesting pod/router-default-5444994796-h9kpk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 08:42:50 crc kubenswrapper[4873]: [-]has-synced failed: reason withheld Dec 01 08:42:50 crc kubenswrapper[4873]: [+]process-running ok Dec 01 08:42:50 crc kubenswrapper[4873]: healthz check failed Dec 01 08:42:50 crc kubenswrapper[4873]: I1201 08:42:50.336311 4873 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h9kpk" podUID="e686de65-3d86-4826-9f0a-4132c81e9ed2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 08:42:50 crc kubenswrapper[4873]: I1201 08:42:50.434834 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:42:50 crc kubenswrapper[4873]: I1201 08:42:50.435430 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:42:50 crc kubenswrapper[4873]: I1201 08:42:50.435467 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:42:50 crc kubenswrapper[4873]: I1201 08:42:50.435490 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:42:50 crc kubenswrapper[4873]: I1201 08:42:50.435524 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:50 crc kubenswrapper[4873]: E1201 08:42:50.435901 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:50.935885344 +0000 UTC m=+146.837993883 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:50 crc kubenswrapper[4873]: I1201 08:42:50.436876 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:42:50 crc kubenswrapper[4873]: I1201 08:42:50.461297 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:42:50 crc kubenswrapper[4873]: I1201 08:42:50.466964 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:42:50 crc kubenswrapper[4873]: I1201 08:42:50.468092 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:42:50 crc kubenswrapper[4873]: I1201 08:42:50.468230 4873 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-f5rc8 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:5443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 01 08:42:50 crc kubenswrapper[4873]: I1201 08:42:50.468264 4873 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f5rc8" podUID="90dd16fd-d069-4007-8c82-4811cb343380" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.32:5443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 01 08:42:50 crc kubenswrapper[4873]: I1201 08:42:50.502071 4873 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-7hstm container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Dec 01 08:42:50 crc kubenswrapper[4873]: I1201 08:42:50.502127 4873 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-7hstm" podUID="2ea695f7-4e9e-4b12-8e0c-22431291576d" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" Dec 01 08:42:50 crc kubenswrapper[4873]: I1201 08:42:50.521350 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-whl8s" Dec 01 08:42:50 crc kubenswrapper[4873]: I1201 08:42:50.605160 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:42:50 crc kubenswrapper[4873]: I1201 08:42:50.607225 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:42:50 crc kubenswrapper[4873]: I1201 08:42:50.607861 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:42:50 crc kubenswrapper[4873]: I1201 08:42:50.607878 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:50 crc kubenswrapper[4873]: E1201 08:42:50.608258 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:51.108236955 +0000 UTC m=+147.010345494 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:50 crc kubenswrapper[4873]: I1201 08:42:50.711336 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:50 crc kubenswrapper[4873]: E1201 08:42:50.714555 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:51.21453842 +0000 UTC m=+147.116646959 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:50 crc kubenswrapper[4873]: I1201 08:42:50.813060 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:50 crc kubenswrapper[4873]: E1201 08:42:50.825288 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:51.313469759 +0000 UTC m=+147.215578298 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:50 crc kubenswrapper[4873]: I1201 08:42:50.838024 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-w8xn2" Dec 01 08:42:50 crc kubenswrapper[4873]: I1201 08:42:50.838066 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-w8xn2" Dec 01 08:42:50 crc kubenswrapper[4873]: I1201 08:42:50.890190 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 01 08:42:50 crc kubenswrapper[4873]: I1201 08:42:50.890818 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 08:42:50 crc kubenswrapper[4873]: I1201 08:42:50.908966 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 01 08:42:50 crc kubenswrapper[4873]: I1201 08:42:50.909266 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 01 08:42:50 crc kubenswrapper[4873]: I1201 08:42:50.915426 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:50 crc kubenswrapper[4873]: E1201 08:42:50.915770 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:51.415756268 +0000 UTC m=+147.317864807 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:50 crc kubenswrapper[4873]: I1201 08:42:50.916141 4873 patch_prober.go:28] interesting pod/apiserver-76f77b778f-w8xn2 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 01 08:42:50 crc kubenswrapper[4873]: [+]log ok Dec 01 08:42:50 crc kubenswrapper[4873]: [+]etcd ok Dec 01 08:42:50 crc kubenswrapper[4873]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 01 08:42:50 crc kubenswrapper[4873]: [+]poststarthook/generic-apiserver-start-informers ok Dec 01 08:42:50 crc kubenswrapper[4873]: [+]poststarthook/max-in-flight-filter ok Dec 01 08:42:50 crc kubenswrapper[4873]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 01 08:42:50 crc kubenswrapper[4873]: [+]poststarthook/image.openshift.io-apiserver-caches ok Dec 01 08:42:50 crc kubenswrapper[4873]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Dec 01 08:42:50 crc kubenswrapper[4873]: [+]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa ok Dec 01 08:42:50 crc kubenswrapper[4873]: [+]poststarthook/project.openshift.io-projectcache ok Dec 01 08:42:50 crc kubenswrapper[4873]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Dec 01 08:42:50 crc kubenswrapper[4873]: [+]poststarthook/openshift.io-startinformers ok Dec 01 08:42:50 crc kubenswrapper[4873]: [+]poststarthook/openshift.io-restmapperupdater ok Dec 01 08:42:50 crc kubenswrapper[4873]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 01 08:42:50 crc kubenswrapper[4873]: livez check failed Dec 01 08:42:50 crc kubenswrapper[4873]: I1201 08:42:50.916222 4873 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-w8xn2" podUID="d0c705dd-9608-4b5e-8c91-2511d8427419" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 08:42:50 crc kubenswrapper[4873]: I1201 08:42:50.936156 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 01 08:42:50 crc kubenswrapper[4873]: I1201 08:42:50.992123 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-292nl" Dec 01 08:42:50 crc kubenswrapper[4873]: I1201 08:42:50.992703 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-292nl" Dec 01 08:42:51 crc kubenswrapper[4873]: I1201 08:42:51.026668 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:51 crc kubenswrapper[4873]: I1201 08:42:51.026920 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b598cb89-a739-44ff-9a29-6e3da327b52a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"b598cb89-a739-44ff-9a29-6e3da327b52a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 08:42:51 crc kubenswrapper[4873]: I1201 08:42:51.027081 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b598cb89-a739-44ff-9a29-6e3da327b52a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"b598cb89-a739-44ff-9a29-6e3da327b52a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 08:42:51 crc kubenswrapper[4873]: E1201 08:42:51.028281 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:51.528228759 +0000 UTC m=+147.430337298 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:51 crc kubenswrapper[4873]: I1201 08:42:51.052462 4873 patch_prober.go:28] interesting pod/console-f9d7485db-292nl container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.28:8443/health\": dial tcp 10.217.0.28:8443: connect: connection refused" start-of-body= Dec 01 08:42:51 crc kubenswrapper[4873]: I1201 08:42:51.052557 4873 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-292nl" podUID="9491cb1f-acf6-438b-8175-11737d7bd245" containerName="console" probeResult="failure" output="Get \"https://10.217.0.28:8443/health\": dial tcp 10.217.0.28:8443: connect: connection refused" Dec 01 08:42:51 crc kubenswrapper[4873]: I1201 08:42:51.128916 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:51 crc kubenswrapper[4873]: I1201 08:42:51.128991 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b598cb89-a739-44ff-9a29-6e3da327b52a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"b598cb89-a739-44ff-9a29-6e3da327b52a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 08:42:51 crc kubenswrapper[4873]: I1201 08:42:51.129075 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b598cb89-a739-44ff-9a29-6e3da327b52a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"b598cb89-a739-44ff-9a29-6e3da327b52a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 08:42:51 crc kubenswrapper[4873]: E1201 08:42:51.131214 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:51.631199915 +0000 UTC m=+147.533308454 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:51 crc kubenswrapper[4873]: I1201 08:42:51.132172 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b598cb89-a739-44ff-9a29-6e3da327b52a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"b598cb89-a739-44ff-9a29-6e3da327b52a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 08:42:51 crc kubenswrapper[4873]: I1201 08:42:51.174178 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b598cb89-a739-44ff-9a29-6e3da327b52a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"b598cb89-a739-44ff-9a29-6e3da327b52a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 08:42:51 crc kubenswrapper[4873]: I1201 08:42:51.227411 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f5rc8" Dec 01 08:42:51 crc kubenswrapper[4873]: I1201 08:42:51.234623 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:51 crc kubenswrapper[4873]: E1201 08:42:51.234927 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:51.734907973 +0000 UTC m=+147.637016512 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:51 crc kubenswrapper[4873]: I1201 08:42:51.334184 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-h9kpk" Dec 01 08:42:51 crc kubenswrapper[4873]: I1201 08:42:51.335487 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:51 crc kubenswrapper[4873]: E1201 08:42:51.335859 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:51.835846075 +0000 UTC m=+147.737954614 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:51 crc kubenswrapper[4873]: I1201 08:42:51.344345 4873 patch_prober.go:28] interesting pod/router-default-5444994796-h9kpk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 08:42:51 crc kubenswrapper[4873]: [-]has-synced failed: reason withheld Dec 01 08:42:51 crc kubenswrapper[4873]: [+]process-running ok Dec 01 08:42:51 crc kubenswrapper[4873]: healthz check failed Dec 01 08:42:51 crc kubenswrapper[4873]: I1201 08:42:51.344412 4873 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h9kpk" podUID="e686de65-3d86-4826-9f0a-4132c81e9ed2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 08:42:51 crc kubenswrapper[4873]: I1201 08:42:51.365078 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 08:42:51 crc kubenswrapper[4873]: I1201 08:42:51.506811 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:51 crc kubenswrapper[4873]: E1201 08:42:51.508144 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:52.008124744 +0000 UTC m=+147.910233283 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:51 crc kubenswrapper[4873]: I1201 08:42:51.511102 4873 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-sg6bg container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.21:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 01 08:42:51 crc kubenswrapper[4873]: I1201 08:42:51.511154 4873 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" podUID="2404e1f8-203a-4055-8f66-21d8867e7a92" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.21:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 01 08:42:51 crc kubenswrapper[4873]: I1201 08:42:51.589428 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9nmnq" event={"ID":"04b9f090-d8f6-4122-bbff-af681fcc4f42","Type":"ContainerStarted","Data":"ca9006e961620ddb969c0aee61ee473f9504148ab9ec8212efcbdea53e4450ec"} Dec 01 08:42:51 crc kubenswrapper[4873]: I1201 08:42:51.609289 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:51 crc kubenswrapper[4873]: E1201 08:42:51.609686 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:52.109672722 +0000 UTC m=+148.011781261 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:51 crc kubenswrapper[4873]: I1201 08:42:51.624906 4873 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-7hstm container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Dec 01 08:42:51 crc kubenswrapper[4873]: I1201 08:42:51.624963 4873 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-7hstm" podUID="2ea695f7-4e9e-4b12-8e0c-22431291576d" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" Dec 01 08:42:51 crc kubenswrapper[4873]: I1201 08:42:51.625055 4873 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-7hstm container/marketplace-operator namespace/openshift-marketplace: Liveness probe status=failure output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Dec 01 08:42:51 crc kubenswrapper[4873]: I1201 08:42:51.625076 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/marketplace-operator-79b997595-7hstm" podUID="2ea695f7-4e9e-4b12-8e0c-22431291576d" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" Dec 01 08:42:51 crc kubenswrapper[4873]: I1201 08:42:51.710880 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:51 crc kubenswrapper[4873]: E1201 08:42:51.711345 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:52.211325144 +0000 UTC m=+148.113433683 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:51 crc kubenswrapper[4873]: I1201 08:42:51.772067 4873 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 01 08:42:51 crc kubenswrapper[4873]: I1201 08:42:51.809199 4873 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-01T08:42:51.772371053Z","Handler":null,"Name":""} Dec 01 08:42:51 crc kubenswrapper[4873]: I1201 08:42:51.816658 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:51 crc kubenswrapper[4873]: E1201 08:42:51.817171 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:52.317155469 +0000 UTC m=+148.219264008 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:51 crc kubenswrapper[4873]: I1201 08:42:51.924564 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:51 crc kubenswrapper[4873]: E1201 08:42:51.924882 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:52.424858272 +0000 UTC m=+148.326966811 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:52 crc kubenswrapper[4873]: I1201 08:42:52.020005 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:42:52 crc kubenswrapper[4873]: I1201 08:42:52.025824 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:52 crc kubenswrapper[4873]: E1201 08:42:52.041402 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:52.54137183 +0000 UTC m=+148.443480369 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cjn5c" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:52 crc kubenswrapper[4873]: I1201 08:42:52.094010 4873 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 01 08:42:52 crc kubenswrapper[4873]: I1201 08:42:52.094095 4873 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 01 08:42:52 crc kubenswrapper[4873]: I1201 08:42:52.128471 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:52 crc kubenswrapper[4873]: I1201 08:42:52.157497 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 01 08:42:52 crc kubenswrapper[4873]: I1201 08:42:52.232783 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:52 crc kubenswrapper[4873]: I1201 08:42:52.288051 4873 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 01 08:42:52 crc kubenswrapper[4873]: I1201 08:42:52.288094 4873 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:52 crc kubenswrapper[4873]: I1201 08:42:52.350218 4873 patch_prober.go:28] interesting pod/router-default-5444994796-h9kpk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 08:42:52 crc kubenswrapper[4873]: [-]has-synced failed: reason withheld Dec 01 08:42:52 crc kubenswrapper[4873]: [+]process-running ok Dec 01 08:42:52 crc kubenswrapper[4873]: healthz check failed Dec 01 08:42:52 crc kubenswrapper[4873]: I1201 08:42:52.350283 4873 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h9kpk" podUID="e686de65-3d86-4826-9f0a-4132c81e9ed2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 08:42:52 crc kubenswrapper[4873]: I1201 08:42:52.438865 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cjn5c\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:52 crc kubenswrapper[4873]: I1201 08:42:52.444228 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 01 08:42:52 crc kubenswrapper[4873]: I1201 08:42:52.567196 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 01 08:42:52 crc kubenswrapper[4873]: I1201 08:42:52.591671 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jr8nt"] Dec 01 08:42:52 crc kubenswrapper[4873]: I1201 08:42:52.593198 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jr8nt" Dec 01 08:42:52 crc kubenswrapper[4873]: I1201 08:42:52.595523 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 01 08:42:52 crc kubenswrapper[4873]: I1201 08:42:52.598873 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"25f843e9748c7a75764b45ddfd53c941ea8400d94c3d8015746c81cb5bc9cc36"} Dec 01 08:42:52 crc kubenswrapper[4873]: I1201 08:42:52.616553 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jr8nt"] Dec 01 08:42:52 crc kubenswrapper[4873]: I1201 08:42:52.618244 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"3f3db0ebef61de21f1a1e4fd7e754def837e56bc1d484a693ad089ac8a1adee2"} Dec 01 08:42:52 crc kubenswrapper[4873]: I1201 08:42:52.641610 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9nmnq" event={"ID":"04b9f090-d8f6-4122-bbff-af681fcc4f42","Type":"ContainerStarted","Data":"bd707279614a3fedc3c83c163b35d52cae191d9925d946ceba39ef783233f8e5"} Dec 01 08:42:52 crc kubenswrapper[4873]: I1201 08:42:52.707555 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:42:52 crc kubenswrapper[4873]: W1201 08:42:52.734025 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-a6215a59e2d724c5439b24b7c069f418b2b3390511cadf24da5db3944ee7f260 WatchSource:0}: Error finding container a6215a59e2d724c5439b24b7c069f418b2b3390511cadf24da5db3944ee7f260: Status 404 returned error can't find the container with id a6215a59e2d724c5439b24b7c069f418b2b3390511cadf24da5db3944ee7f260 Dec 01 08:42:52 crc kubenswrapper[4873]: I1201 08:42:52.734463 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:52 crc kubenswrapper[4873]: I1201 08:42:52.755059 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d44d4877-74e9-4fdc-a062-92d031f1636d-catalog-content\") pod \"certified-operators-jr8nt\" (UID: \"d44d4877-74e9-4fdc-a062-92d031f1636d\") " pod="openshift-marketplace/certified-operators-jr8nt" Dec 01 08:42:52 crc kubenswrapper[4873]: I1201 08:42:52.755118 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gg4z6\" (UniqueName: \"kubernetes.io/projected/d44d4877-74e9-4fdc-a062-92d031f1636d-kube-api-access-gg4z6\") pod \"certified-operators-jr8nt\" (UID: \"d44d4877-74e9-4fdc-a062-92d031f1636d\") " pod="openshift-marketplace/certified-operators-jr8nt" Dec 01 08:42:52 crc kubenswrapper[4873]: I1201 08:42:52.755170 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d44d4877-74e9-4fdc-a062-92d031f1636d-utilities\") pod \"certified-operators-jr8nt\" (UID: \"d44d4877-74e9-4fdc-a062-92d031f1636d\") " pod="openshift-marketplace/certified-operators-jr8nt" Dec 01 08:42:52 crc kubenswrapper[4873]: I1201 08:42:52.844740 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-dm6vl"] Dec 01 08:42:52 crc kubenswrapper[4873]: I1201 08:42:52.850277 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dm6vl" Dec 01 08:42:52 crc kubenswrapper[4873]: I1201 08:42:52.858537 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d44d4877-74e9-4fdc-a062-92d031f1636d-catalog-content\") pod \"certified-operators-jr8nt\" (UID: \"d44d4877-74e9-4fdc-a062-92d031f1636d\") " pod="openshift-marketplace/certified-operators-jr8nt" Dec 01 08:42:52 crc kubenswrapper[4873]: I1201 08:42:52.858623 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gg4z6\" (UniqueName: \"kubernetes.io/projected/d44d4877-74e9-4fdc-a062-92d031f1636d-kube-api-access-gg4z6\") pod \"certified-operators-jr8nt\" (UID: \"d44d4877-74e9-4fdc-a062-92d031f1636d\") " pod="openshift-marketplace/certified-operators-jr8nt" Dec 01 08:42:52 crc kubenswrapper[4873]: I1201 08:42:52.859307 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d44d4877-74e9-4fdc-a062-92d031f1636d-utilities\") pod \"certified-operators-jr8nt\" (UID: \"d44d4877-74e9-4fdc-a062-92d031f1636d\") " pod="openshift-marketplace/certified-operators-jr8nt" Dec 01 08:42:52 crc kubenswrapper[4873]: I1201 08:42:52.859312 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d44d4877-74e9-4fdc-a062-92d031f1636d-catalog-content\") pod \"certified-operators-jr8nt\" (UID: \"d44d4877-74e9-4fdc-a062-92d031f1636d\") " pod="openshift-marketplace/certified-operators-jr8nt" Dec 01 08:42:52 crc kubenswrapper[4873]: I1201 08:42:52.859546 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d44d4877-74e9-4fdc-a062-92d031f1636d-utilities\") pod \"certified-operators-jr8nt\" (UID: \"d44d4877-74e9-4fdc-a062-92d031f1636d\") " pod="openshift-marketplace/certified-operators-jr8nt" Dec 01 08:42:52 crc kubenswrapper[4873]: I1201 08:42:52.871512 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 01 08:42:52 crc kubenswrapper[4873]: I1201 08:42:52.901461 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gg4z6\" (UniqueName: \"kubernetes.io/projected/d44d4877-74e9-4fdc-a062-92d031f1636d-kube-api-access-gg4z6\") pod \"certified-operators-jr8nt\" (UID: \"d44d4877-74e9-4fdc-a062-92d031f1636d\") " pod="openshift-marketplace/certified-operators-jr8nt" Dec 01 08:42:52 crc kubenswrapper[4873]: I1201 08:42:52.918712 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dm6vl"] Dec 01 08:42:52 crc kubenswrapper[4873]: I1201 08:42:52.960158 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a18563f-c79c-4fed-83fb-5ee7d865e014-utilities\") pod \"community-operators-dm6vl\" (UID: \"7a18563f-c79c-4fed-83fb-5ee7d865e014\") " pod="openshift-marketplace/community-operators-dm6vl" Dec 01 08:42:52 crc kubenswrapper[4873]: I1201 08:42:52.960294 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a18563f-c79c-4fed-83fb-5ee7d865e014-catalog-content\") pod \"community-operators-dm6vl\" (UID: \"7a18563f-c79c-4fed-83fb-5ee7d865e014\") " pod="openshift-marketplace/community-operators-dm6vl" Dec 01 08:42:52 crc kubenswrapper[4873]: I1201 08:42:52.960353 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pl7wx\" (UniqueName: \"kubernetes.io/projected/7a18563f-c79c-4fed-83fb-5ee7d865e014-kube-api-access-pl7wx\") pod \"community-operators-dm6vl\" (UID: \"7a18563f-c79c-4fed-83fb-5ee7d865e014\") " pod="openshift-marketplace/community-operators-dm6vl" Dec 01 08:42:52 crc kubenswrapper[4873]: I1201 08:42:52.987117 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jr8nt" Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.002614 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-x5tjn"] Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.003775 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x5tjn" Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.040430 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-x5tjn"] Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.061628 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc58a1d4-9469-4205-abb9-ef87eb162644-catalog-content\") pod \"certified-operators-x5tjn\" (UID: \"cc58a1d4-9469-4205-abb9-ef87eb162644\") " pod="openshift-marketplace/certified-operators-x5tjn" Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.061692 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc58a1d4-9469-4205-abb9-ef87eb162644-utilities\") pod \"certified-operators-x5tjn\" (UID: \"cc58a1d4-9469-4205-abb9-ef87eb162644\") " pod="openshift-marketplace/certified-operators-x5tjn" Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.061750 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a18563f-c79c-4fed-83fb-5ee7d865e014-utilities\") pod \"community-operators-dm6vl\" (UID: \"7a18563f-c79c-4fed-83fb-5ee7d865e014\") " pod="openshift-marketplace/community-operators-dm6vl" Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.061818 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a18563f-c79c-4fed-83fb-5ee7d865e014-catalog-content\") pod \"community-operators-dm6vl\" (UID: \"7a18563f-c79c-4fed-83fb-5ee7d865e014\") " pod="openshift-marketplace/community-operators-dm6vl" Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.061854 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgkbq\" (UniqueName: \"kubernetes.io/projected/cc58a1d4-9469-4205-abb9-ef87eb162644-kube-api-access-vgkbq\") pod \"certified-operators-x5tjn\" (UID: \"cc58a1d4-9469-4205-abb9-ef87eb162644\") " pod="openshift-marketplace/certified-operators-x5tjn" Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.061876 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pl7wx\" (UniqueName: \"kubernetes.io/projected/7a18563f-c79c-4fed-83fb-5ee7d865e014-kube-api-access-pl7wx\") pod \"community-operators-dm6vl\" (UID: \"7a18563f-c79c-4fed-83fb-5ee7d865e014\") " pod="openshift-marketplace/community-operators-dm6vl" Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.062798 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a18563f-c79c-4fed-83fb-5ee7d865e014-utilities\") pod \"community-operators-dm6vl\" (UID: \"7a18563f-c79c-4fed-83fb-5ee7d865e014\") " pod="openshift-marketplace/community-operators-dm6vl" Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.063067 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a18563f-c79c-4fed-83fb-5ee7d865e014-catalog-content\") pod \"community-operators-dm6vl\" (UID: \"7a18563f-c79c-4fed-83fb-5ee7d865e014\") " pod="openshift-marketplace/community-operators-dm6vl" Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.127030 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pl7wx\" (UniqueName: \"kubernetes.io/projected/7a18563f-c79c-4fed-83fb-5ee7d865e014-kube-api-access-pl7wx\") pod \"community-operators-dm6vl\" (UID: \"7a18563f-c79c-4fed-83fb-5ee7d865e014\") " pod="openshift-marketplace/community-operators-dm6vl" Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.163790 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgkbq\" (UniqueName: \"kubernetes.io/projected/cc58a1d4-9469-4205-abb9-ef87eb162644-kube-api-access-vgkbq\") pod \"certified-operators-x5tjn\" (UID: \"cc58a1d4-9469-4205-abb9-ef87eb162644\") " pod="openshift-marketplace/certified-operators-x5tjn" Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.163963 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc58a1d4-9469-4205-abb9-ef87eb162644-catalog-content\") pod \"certified-operators-x5tjn\" (UID: \"cc58a1d4-9469-4205-abb9-ef87eb162644\") " pod="openshift-marketplace/certified-operators-x5tjn" Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.163989 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc58a1d4-9469-4205-abb9-ef87eb162644-utilities\") pod \"certified-operators-x5tjn\" (UID: \"cc58a1d4-9469-4205-abb9-ef87eb162644\") " pod="openshift-marketplace/certified-operators-x5tjn" Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.164829 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc58a1d4-9469-4205-abb9-ef87eb162644-utilities\") pod \"certified-operators-x5tjn\" (UID: \"cc58a1d4-9469-4205-abb9-ef87eb162644\") " pod="openshift-marketplace/certified-operators-x5tjn" Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.165473 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc58a1d4-9469-4205-abb9-ef87eb162644-catalog-content\") pod \"certified-operators-x5tjn\" (UID: \"cc58a1d4-9469-4205-abb9-ef87eb162644\") " pod="openshift-marketplace/certified-operators-x5tjn" Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.172462 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8nchr"] Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.176384 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8nchr" Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.207127 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dm6vl" Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.207604 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgkbq\" (UniqueName: \"kubernetes.io/projected/cc58a1d4-9469-4205-abb9-ef87eb162644-kube-api-access-vgkbq\") pod \"certified-operators-x5tjn\" (UID: \"cc58a1d4-9469-4205-abb9-ef87eb162644\") " pod="openshift-marketplace/certified-operators-x5tjn" Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.212569 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8nchr"] Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.268627 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d7debb7-93da-4965-9fb9-fe8626b68183-catalog-content\") pod \"community-operators-8nchr\" (UID: \"1d7debb7-93da-4965-9fb9-fe8626b68183\") " pod="openshift-marketplace/community-operators-8nchr" Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.269224 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d7debb7-93da-4965-9fb9-fe8626b68183-utilities\") pod \"community-operators-8nchr\" (UID: \"1d7debb7-93da-4965-9fb9-fe8626b68183\") " pod="openshift-marketplace/community-operators-8nchr" Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.269303 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67mng\" (UniqueName: \"kubernetes.io/projected/1d7debb7-93da-4965-9fb9-fe8626b68183-kube-api-access-67mng\") pod \"community-operators-8nchr\" (UID: \"1d7debb7-93da-4965-9fb9-fe8626b68183\") " pod="openshift-marketplace/community-operators-8nchr" Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.340439 4873 patch_prober.go:28] interesting pod/router-default-5444994796-h9kpk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 08:42:53 crc kubenswrapper[4873]: [-]has-synced failed: reason withheld Dec 01 08:42:53 crc kubenswrapper[4873]: [+]process-running ok Dec 01 08:42:53 crc kubenswrapper[4873]: healthz check failed Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.340517 4873 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h9kpk" podUID="e686de65-3d86-4826-9f0a-4132c81e9ed2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.368032 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cjn5c"] Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.370900 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67mng\" (UniqueName: \"kubernetes.io/projected/1d7debb7-93da-4965-9fb9-fe8626b68183-kube-api-access-67mng\") pod \"community-operators-8nchr\" (UID: \"1d7debb7-93da-4965-9fb9-fe8626b68183\") " pod="openshift-marketplace/community-operators-8nchr" Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.370987 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d7debb7-93da-4965-9fb9-fe8626b68183-catalog-content\") pod \"community-operators-8nchr\" (UID: \"1d7debb7-93da-4965-9fb9-fe8626b68183\") " pod="openshift-marketplace/community-operators-8nchr" Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.371061 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d7debb7-93da-4965-9fb9-fe8626b68183-utilities\") pod \"community-operators-8nchr\" (UID: \"1d7debb7-93da-4965-9fb9-fe8626b68183\") " pod="openshift-marketplace/community-operators-8nchr" Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.371705 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d7debb7-93da-4965-9fb9-fe8626b68183-utilities\") pod \"community-operators-8nchr\" (UID: \"1d7debb7-93da-4965-9fb9-fe8626b68183\") " pod="openshift-marketplace/community-operators-8nchr" Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.371807 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d7debb7-93da-4965-9fb9-fe8626b68183-catalog-content\") pod \"community-operators-8nchr\" (UID: \"1d7debb7-93da-4965-9fb9-fe8626b68183\") " pod="openshift-marketplace/community-operators-8nchr" Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.381631 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x5tjn" Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.392878 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67mng\" (UniqueName: \"kubernetes.io/projected/1d7debb7-93da-4965-9fb9-fe8626b68183-kube-api-access-67mng\") pod \"community-operators-8nchr\" (UID: \"1d7debb7-93da-4965-9fb9-fe8626b68183\") " pod="openshift-marketplace/community-operators-8nchr" Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.502894 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jr8nt"] Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.524408 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8nchr" Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.594953 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dm6vl"] Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.672140 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"c16e424a25e166fbcc226b21d9c5e0aa88dc233592ae2b05ceca6ca890c9c94f"} Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.672606 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"a6215a59e2d724c5439b24b7c069f418b2b3390511cadf24da5db3944ee7f260"} Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.720332 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"3e9aa878e61025733e57b2e39216bb5546ae4aab9535fa9711c480f6b885006e"} Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.726435 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jr8nt" event={"ID":"d44d4877-74e9-4fdc-a062-92d031f1636d","Type":"ContainerStarted","Data":"0f0ac101db567bf9028ba04cc6e1fb6fd0d75b2b1e6ac471cd2df8eb0ebf7b9f"} Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.729650 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9nmnq" event={"ID":"04b9f090-d8f6-4122-bbff-af681fcc4f42","Type":"ContainerStarted","Data":"b86c6c482ae6382dd80d7f789b483c0b38febf32e13de2adf7680e48a0d635a8"} Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.730617 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"64d8d7fb01a14e58672d1d475957d07678e8421efc31320f86f8adf64fa6645c"} Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.731352 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.736708 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"b598cb89-a739-44ff-9a29-6e3da327b52a","Type":"ContainerStarted","Data":"1d66897bee5e539bd1687e8b1dfc07426aecf2c6e8c63d362838e240341905ab"} Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.736753 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"b598cb89-a739-44ff-9a29-6e3da327b52a","Type":"ContainerStarted","Data":"cdf61c7e9faed9b6b580aba2eb576bd092a47e9e427e4de3740dd7c262967b1c"} Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.746666 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" event={"ID":"952ae5ba-156f-41a7-950a-1783e2e69f94","Type":"ContainerStarted","Data":"dc0be1a86597c49c7588c707a48263eb632a20b62d5ad5675cf32e8dfe524f9b"} Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.749009 4873 generic.go:334] "Generic (PLEG): container finished" podID="83688ccb-ace0-4b0a-b584-e650671f5686" containerID="38b8fc7cc310a3ec72fe0ed1ba4dbfa51153edeb1fc2eb5f97423a83c30f7bc2" exitCode=0 Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.749076 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-lnw2h" event={"ID":"83688ccb-ace0-4b0a-b584-e650671f5686","Type":"ContainerDied","Data":"38b8fc7cc310a3ec72fe0ed1ba4dbfa51153edeb1fc2eb5f97423a83c30f7bc2"} Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.777802 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-x5tjn"] Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.802709 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-9nmnq" podStartSLOduration=15.802677401 podStartE2EDuration="15.802677401s" podCreationTimestamp="2025-12-01 08:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:53.789568928 +0000 UTC m=+149.691677477" watchObservedRunningTime="2025-12-01 08:42:53.802677401 +0000 UTC m=+149.704785940" Dec 01 08:42:53 crc kubenswrapper[4873]: I1201 08:42:53.807922 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=3.807898538 podStartE2EDuration="3.807898538s" podCreationTimestamp="2025-12-01 08:42:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:53.806762981 +0000 UTC m=+149.708871520" watchObservedRunningTime="2025-12-01 08:42:53.807898538 +0000 UTC m=+149.710007077" Dec 01 08:42:54 crc kubenswrapper[4873]: I1201 08:42:54.094799 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8nchr"] Dec 01 08:42:54 crc kubenswrapper[4873]: I1201 08:42:54.337379 4873 patch_prober.go:28] interesting pod/router-default-5444994796-h9kpk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 08:42:54 crc kubenswrapper[4873]: [-]has-synced failed: reason withheld Dec 01 08:42:54 crc kubenswrapper[4873]: [+]process-running ok Dec 01 08:42:54 crc kubenswrapper[4873]: healthz check failed Dec 01 08:42:54 crc kubenswrapper[4873]: I1201 08:42:54.337799 4873 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h9kpk" podUID="e686de65-3d86-4826-9f0a-4132c81e9ed2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 08:42:54 crc kubenswrapper[4873]: I1201 08:42:54.587663 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-spckc"] Dec 01 08:42:54 crc kubenswrapper[4873]: I1201 08:42:54.589324 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-spckc" Dec 01 08:42:54 crc kubenswrapper[4873]: I1201 08:42:54.590938 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 01 08:42:54 crc kubenswrapper[4873]: I1201 08:42:54.603491 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-spckc"] Dec 01 08:42:54 crc kubenswrapper[4873]: I1201 08:42:54.695377 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73df40c6-5524-496f-92ba-36100c4af8bb-catalog-content\") pod \"redhat-marketplace-spckc\" (UID: \"73df40c6-5524-496f-92ba-36100c4af8bb\") " pod="openshift-marketplace/redhat-marketplace-spckc" Dec 01 08:42:54 crc kubenswrapper[4873]: I1201 08:42:54.695590 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flxnb\" (UniqueName: \"kubernetes.io/projected/73df40c6-5524-496f-92ba-36100c4af8bb-kube-api-access-flxnb\") pod \"redhat-marketplace-spckc\" (UID: \"73df40c6-5524-496f-92ba-36100c4af8bb\") " pod="openshift-marketplace/redhat-marketplace-spckc" Dec 01 08:42:54 crc kubenswrapper[4873]: I1201 08:42:54.695670 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73df40c6-5524-496f-92ba-36100c4af8bb-utilities\") pod \"redhat-marketplace-spckc\" (UID: \"73df40c6-5524-496f-92ba-36100c4af8bb\") " pod="openshift-marketplace/redhat-marketplace-spckc" Dec 01 08:42:54 crc kubenswrapper[4873]: I1201 08:42:54.756137 4873 generic.go:334] "Generic (PLEG): container finished" podID="b598cb89-a739-44ff-9a29-6e3da327b52a" containerID="1d66897bee5e539bd1687e8b1dfc07426aecf2c6e8c63d362838e240341905ab" exitCode=0 Dec 01 08:42:54 crc kubenswrapper[4873]: I1201 08:42:54.756215 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"b598cb89-a739-44ff-9a29-6e3da327b52a","Type":"ContainerDied","Data":"1d66897bee5e539bd1687e8b1dfc07426aecf2c6e8c63d362838e240341905ab"} Dec 01 08:42:54 crc kubenswrapper[4873]: I1201 08:42:54.758980 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" event={"ID":"952ae5ba-156f-41a7-950a-1783e2e69f94","Type":"ContainerStarted","Data":"be35e87e40c6a12a9a8bf6f80153adc4d38b8d8466027397bc51f10e7db3f1c0"} Dec 01 08:42:54 crc kubenswrapper[4873]: I1201 08:42:54.759777 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:42:54 crc kubenswrapper[4873]: I1201 08:42:54.763451 4873 generic.go:334] "Generic (PLEG): container finished" podID="7a18563f-c79c-4fed-83fb-5ee7d865e014" containerID="66fcda06001cbdaccd70eea2d66862bc764bf489bf2ffe5da590e6223e8109dd" exitCode=0 Dec 01 08:42:54 crc kubenswrapper[4873]: I1201 08:42:54.763586 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dm6vl" event={"ID":"7a18563f-c79c-4fed-83fb-5ee7d865e014","Type":"ContainerDied","Data":"66fcda06001cbdaccd70eea2d66862bc764bf489bf2ffe5da590e6223e8109dd"} Dec 01 08:42:54 crc kubenswrapper[4873]: I1201 08:42:54.764038 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dm6vl" event={"ID":"7a18563f-c79c-4fed-83fb-5ee7d865e014","Type":"ContainerStarted","Data":"811c5560bd438367a21c284150b9c460ab798fe6be675bdc68322faf6c604744"} Dec 01 08:42:54 crc kubenswrapper[4873]: I1201 08:42:54.765416 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8nchr" event={"ID":"1d7debb7-93da-4965-9fb9-fe8626b68183","Type":"ContainerStarted","Data":"ccbccaac54d39aeb561dbebb217e4073d3f172fc70459c35f8baec69f5724f8f"} Dec 01 08:42:54 crc kubenswrapper[4873]: I1201 08:42:54.767077 4873 generic.go:334] "Generic (PLEG): container finished" podID="cc58a1d4-9469-4205-abb9-ef87eb162644" containerID="b9ca40c29ce38d12a47b10dec6ee4674c41cdeae2a58f13b83d3db551c095444" exitCode=0 Dec 01 08:42:54 crc kubenswrapper[4873]: I1201 08:42:54.767137 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x5tjn" event={"ID":"cc58a1d4-9469-4205-abb9-ef87eb162644","Type":"ContainerDied","Data":"b9ca40c29ce38d12a47b10dec6ee4674c41cdeae2a58f13b83d3db551c095444"} Dec 01 08:42:54 crc kubenswrapper[4873]: I1201 08:42:54.767157 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x5tjn" event={"ID":"cc58a1d4-9469-4205-abb9-ef87eb162644","Type":"ContainerStarted","Data":"2516f86433ffd5a73c2b5bb06e7d5d25c158806b10e8b33d50ae55e2eabe2283"} Dec 01 08:42:54 crc kubenswrapper[4873]: I1201 08:42:54.771859 4873 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 08:42:54 crc kubenswrapper[4873]: I1201 08:42:54.778594 4873 generic.go:334] "Generic (PLEG): container finished" podID="d44d4877-74e9-4fdc-a062-92d031f1636d" containerID="76bf7b52eeff1ac99c60219084b867f2cd1abb47d6fea23a99461dea51d606a5" exitCode=0 Dec 01 08:42:54 crc kubenswrapper[4873]: I1201 08:42:54.779305 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jr8nt" event={"ID":"d44d4877-74e9-4fdc-a062-92d031f1636d","Type":"ContainerDied","Data":"76bf7b52eeff1ac99c60219084b867f2cd1abb47d6fea23a99461dea51d606a5"} Dec 01 08:42:54 crc kubenswrapper[4873]: I1201 08:42:54.799730 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73df40c6-5524-496f-92ba-36100c4af8bb-utilities\") pod \"redhat-marketplace-spckc\" (UID: \"73df40c6-5524-496f-92ba-36100c4af8bb\") " pod="openshift-marketplace/redhat-marketplace-spckc" Dec 01 08:42:54 crc kubenswrapper[4873]: I1201 08:42:54.799819 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73df40c6-5524-496f-92ba-36100c4af8bb-catalog-content\") pod \"redhat-marketplace-spckc\" (UID: \"73df40c6-5524-496f-92ba-36100c4af8bb\") " pod="openshift-marketplace/redhat-marketplace-spckc" Dec 01 08:42:54 crc kubenswrapper[4873]: I1201 08:42:54.799891 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flxnb\" (UniqueName: \"kubernetes.io/projected/73df40c6-5524-496f-92ba-36100c4af8bb-kube-api-access-flxnb\") pod \"redhat-marketplace-spckc\" (UID: \"73df40c6-5524-496f-92ba-36100c4af8bb\") " pod="openshift-marketplace/redhat-marketplace-spckc" Dec 01 08:42:54 crc kubenswrapper[4873]: I1201 08:42:54.801235 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73df40c6-5524-496f-92ba-36100c4af8bb-utilities\") pod \"redhat-marketplace-spckc\" (UID: \"73df40c6-5524-496f-92ba-36100c4af8bb\") " pod="openshift-marketplace/redhat-marketplace-spckc" Dec 01 08:42:54 crc kubenswrapper[4873]: I1201 08:42:54.802196 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73df40c6-5524-496f-92ba-36100c4af8bb-catalog-content\") pod \"redhat-marketplace-spckc\" (UID: \"73df40c6-5524-496f-92ba-36100c4af8bb\") " pod="openshift-marketplace/redhat-marketplace-spckc" Dec 01 08:42:54 crc kubenswrapper[4873]: I1201 08:42:54.842828 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flxnb\" (UniqueName: \"kubernetes.io/projected/73df40c6-5524-496f-92ba-36100c4af8bb-kube-api-access-flxnb\") pod \"redhat-marketplace-spckc\" (UID: \"73df40c6-5524-496f-92ba-36100c4af8bb\") " pod="openshift-marketplace/redhat-marketplace-spckc" Dec 01 08:42:54 crc kubenswrapper[4873]: I1201 08:42:54.930535 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" podStartSLOduration=131.930505785 podStartE2EDuration="2m11.930505785s" podCreationTimestamp="2025-12-01 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:54.926674666 +0000 UTC m=+150.828783205" watchObservedRunningTime="2025-12-01 08:42:54.930505785 +0000 UTC m=+150.832614324" Dec 01 08:42:54 crc kubenswrapper[4873]: I1201 08:42:54.986515 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-q7s8h"] Dec 01 08:42:54 crc kubenswrapper[4873]: I1201 08:42:54.993931 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q7s8h" Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.007970 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q7s8h"] Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.049766 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-spckc" Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.103495 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-px8th\" (UniqueName: \"kubernetes.io/projected/8c34ba9e-07db-478a-8a30-8622f4de4828-kube-api-access-px8th\") pod \"redhat-marketplace-q7s8h\" (UID: \"8c34ba9e-07db-478a-8a30-8622f4de4828\") " pod="openshift-marketplace/redhat-marketplace-q7s8h" Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.103575 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c34ba9e-07db-478a-8a30-8622f4de4828-catalog-content\") pod \"redhat-marketplace-q7s8h\" (UID: \"8c34ba9e-07db-478a-8a30-8622f4de4828\") " pod="openshift-marketplace/redhat-marketplace-q7s8h" Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.103612 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c34ba9e-07db-478a-8a30-8622f4de4828-utilities\") pod \"redhat-marketplace-q7s8h\" (UID: \"8c34ba9e-07db-478a-8a30-8622f4de4828\") " pod="openshift-marketplace/redhat-marketplace-q7s8h" Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.204905 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-px8th\" (UniqueName: \"kubernetes.io/projected/8c34ba9e-07db-478a-8a30-8622f4de4828-kube-api-access-px8th\") pod \"redhat-marketplace-q7s8h\" (UID: \"8c34ba9e-07db-478a-8a30-8622f4de4828\") " pod="openshift-marketplace/redhat-marketplace-q7s8h" Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.205037 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c34ba9e-07db-478a-8a30-8622f4de4828-catalog-content\") pod \"redhat-marketplace-q7s8h\" (UID: \"8c34ba9e-07db-478a-8a30-8622f4de4828\") " pod="openshift-marketplace/redhat-marketplace-q7s8h" Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.205067 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c34ba9e-07db-478a-8a30-8622f4de4828-utilities\") pod \"redhat-marketplace-q7s8h\" (UID: \"8c34ba9e-07db-478a-8a30-8622f4de4828\") " pod="openshift-marketplace/redhat-marketplace-q7s8h" Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.205946 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c34ba9e-07db-478a-8a30-8622f4de4828-utilities\") pod \"redhat-marketplace-q7s8h\" (UID: \"8c34ba9e-07db-478a-8a30-8622f4de4828\") " pod="openshift-marketplace/redhat-marketplace-q7s8h" Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.206106 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c34ba9e-07db-478a-8a30-8622f4de4828-catalog-content\") pod \"redhat-marketplace-q7s8h\" (UID: \"8c34ba9e-07db-478a-8a30-8622f4de4828\") " pod="openshift-marketplace/redhat-marketplace-q7s8h" Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.240884 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-px8th\" (UniqueName: \"kubernetes.io/projected/8c34ba9e-07db-478a-8a30-8622f4de4828-kube-api-access-px8th\") pod \"redhat-marketplace-q7s8h\" (UID: \"8c34ba9e-07db-478a-8a30-8622f4de4828\") " pod="openshift-marketplace/redhat-marketplace-q7s8h" Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.266414 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.267727 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.272264 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.273160 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.275576 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.306065 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/143397c7-7625-40da-941c-0fe2f2ef6e69-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"143397c7-7625-40da-941c-0fe2f2ef6e69\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.306148 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/143397c7-7625-40da-941c-0fe2f2ef6e69-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"143397c7-7625-40da-941c-0fe2f2ef6e69\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.407053 4873 patch_prober.go:28] interesting pod/router-default-5444994796-h9kpk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 08:42:55 crc kubenswrapper[4873]: [-]has-synced failed: reason withheld Dec 01 08:42:55 crc kubenswrapper[4873]: [+]process-running ok Dec 01 08:42:55 crc kubenswrapper[4873]: healthz check failed Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.407144 4873 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h9kpk" podUID="e686de65-3d86-4826-9f0a-4132c81e9ed2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.409240 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/143397c7-7625-40da-941c-0fe2f2ef6e69-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"143397c7-7625-40da-941c-0fe2f2ef6e69\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.409302 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/143397c7-7625-40da-941c-0fe2f2ef6e69-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"143397c7-7625-40da-941c-0fe2f2ef6e69\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.409420 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/143397c7-7625-40da-941c-0fe2f2ef6e69-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"143397c7-7625-40da-941c-0fe2f2ef6e69\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.411590 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q7s8h" Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.438853 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/143397c7-7625-40da-941c-0fe2f2ef6e69-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"143397c7-7625-40da-941c-0fe2f2ef6e69\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.478884 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.640374 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-spckc"] Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.647527 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-lnw2h" Dec 01 08:42:55 crc kubenswrapper[4873]: W1201 08:42:55.669340 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod73df40c6_5524_496f_92ba_36100c4af8bb.slice/crio-e2c9ca91af2103d1045c7daf00e5177e3c887669f99479ad1ca9b091eabc6446 WatchSource:0}: Error finding container e2c9ca91af2103d1045c7daf00e5177e3c887669f99479ad1ca9b091eabc6446: Status 404 returned error can't find the container with id e2c9ca91af2103d1045c7daf00e5177e3c887669f99479ad1ca9b091eabc6446 Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.716843 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-72gzp\" (UniqueName: \"kubernetes.io/projected/83688ccb-ace0-4b0a-b584-e650671f5686-kube-api-access-72gzp\") pod \"83688ccb-ace0-4b0a-b584-e650671f5686\" (UID: \"83688ccb-ace0-4b0a-b584-e650671f5686\") " Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.716959 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/83688ccb-ace0-4b0a-b584-e650671f5686-config-volume\") pod \"83688ccb-ace0-4b0a-b584-e650671f5686\" (UID: \"83688ccb-ace0-4b0a-b584-e650671f5686\") " Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.717001 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/83688ccb-ace0-4b0a-b584-e650671f5686-secret-volume\") pod \"83688ccb-ace0-4b0a-b584-e650671f5686\" (UID: \"83688ccb-ace0-4b0a-b584-e650671f5686\") " Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.719690 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83688ccb-ace0-4b0a-b584-e650671f5686-config-volume" (OuterVolumeSpecName: "config-volume") pod "83688ccb-ace0-4b0a-b584-e650671f5686" (UID: "83688ccb-ace0-4b0a-b584-e650671f5686"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.723930 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83688ccb-ace0-4b0a-b584-e650671f5686-kube-api-access-72gzp" (OuterVolumeSpecName: "kube-api-access-72gzp") pod "83688ccb-ace0-4b0a-b584-e650671f5686" (UID: "83688ccb-ace0-4b0a-b584-e650671f5686"). InnerVolumeSpecName "kube-api-access-72gzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.726880 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83688ccb-ace0-4b0a-b584-e650671f5686-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "83688ccb-ace0-4b0a-b584-e650671f5686" (UID: "83688ccb-ace0-4b0a-b584-e650671f5686"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.754414 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q7s8h"] Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.779850 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zffmt"] Dec 01 08:42:55 crc kubenswrapper[4873]: E1201 08:42:55.781194 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83688ccb-ace0-4b0a-b584-e650671f5686" containerName="collect-profiles" Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.781260 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="83688ccb-ace0-4b0a-b584-e650671f5686" containerName="collect-profiles" Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.781418 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="83688ccb-ace0-4b0a-b584-e650671f5686" containerName="collect-profiles" Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.782251 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zffmt" Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.784428 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.793003 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zffmt"] Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.797508 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-spckc" event={"ID":"73df40c6-5524-496f-92ba-36100c4af8bb","Type":"ContainerStarted","Data":"e2c9ca91af2103d1045c7daf00e5177e3c887669f99479ad1ca9b091eabc6446"} Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.810371 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-lnw2h" event={"ID":"83688ccb-ace0-4b0a-b584-e650671f5686","Type":"ContainerDied","Data":"c0555dcd08053d6bc41d0f8e5a5c15a6164f017939d68122c620aaa57bed825f"} Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.810426 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c0555dcd08053d6bc41d0f8e5a5c15a6164f017939d68122c620aaa57bed825f" Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.810501 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-lnw2h" Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.812299 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q7s8h" event={"ID":"8c34ba9e-07db-478a-8a30-8622f4de4828","Type":"ContainerStarted","Data":"5a9d6a26628fafc169e37a864d580038b77a3e5306de4c57c20c52610e6acb78"} Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.813925 4873 generic.go:334] "Generic (PLEG): container finished" podID="1d7debb7-93da-4965-9fb9-fe8626b68183" containerID="c35383335fd4da5bbd0b644dd42172b196363f22a819a12b5d67d527c2c2331f" exitCode=0 Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.814111 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8nchr" event={"ID":"1d7debb7-93da-4965-9fb9-fe8626b68183","Type":"ContainerDied","Data":"c35383335fd4da5bbd0b644dd42172b196363f22a819a12b5d67d527c2c2331f"} Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.818460 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8s74\" (UniqueName: \"kubernetes.io/projected/a3fc6e9a-1447-41e2-8886-e5ea32e3d353-kube-api-access-b8s74\") pod \"redhat-operators-zffmt\" (UID: \"a3fc6e9a-1447-41e2-8886-e5ea32e3d353\") " pod="openshift-marketplace/redhat-operators-zffmt" Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.818580 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3fc6e9a-1447-41e2-8886-e5ea32e3d353-catalog-content\") pod \"redhat-operators-zffmt\" (UID: \"a3fc6e9a-1447-41e2-8886-e5ea32e3d353\") " pod="openshift-marketplace/redhat-operators-zffmt" Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.818621 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3fc6e9a-1447-41e2-8886-e5ea32e3d353-utilities\") pod \"redhat-operators-zffmt\" (UID: \"a3fc6e9a-1447-41e2-8886-e5ea32e3d353\") " pod="openshift-marketplace/redhat-operators-zffmt" Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.818767 4873 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/83688ccb-ace0-4b0a-b584-e650671f5686-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.818783 4873 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/83688ccb-ace0-4b0a-b584-e650671f5686-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.818794 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-72gzp\" (UniqueName: \"kubernetes.io/projected/83688ccb-ace0-4b0a-b584-e650671f5686-kube-api-access-72gzp\") on node \"crc\" DevicePath \"\"" Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.853875 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-w8xn2" Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.859964 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-w8xn2" Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.867391 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.919910 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8s74\" (UniqueName: \"kubernetes.io/projected/a3fc6e9a-1447-41e2-8886-e5ea32e3d353-kube-api-access-b8s74\") pod \"redhat-operators-zffmt\" (UID: \"a3fc6e9a-1447-41e2-8886-e5ea32e3d353\") " pod="openshift-marketplace/redhat-operators-zffmt" Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.920029 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3fc6e9a-1447-41e2-8886-e5ea32e3d353-catalog-content\") pod \"redhat-operators-zffmt\" (UID: \"a3fc6e9a-1447-41e2-8886-e5ea32e3d353\") " pod="openshift-marketplace/redhat-operators-zffmt" Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.920072 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3fc6e9a-1447-41e2-8886-e5ea32e3d353-utilities\") pod \"redhat-operators-zffmt\" (UID: \"a3fc6e9a-1447-41e2-8886-e5ea32e3d353\") " pod="openshift-marketplace/redhat-operators-zffmt" Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.922220 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3fc6e9a-1447-41e2-8886-e5ea32e3d353-catalog-content\") pod \"redhat-operators-zffmt\" (UID: \"a3fc6e9a-1447-41e2-8886-e5ea32e3d353\") " pod="openshift-marketplace/redhat-operators-zffmt" Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.922465 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3fc6e9a-1447-41e2-8886-e5ea32e3d353-utilities\") pod \"redhat-operators-zffmt\" (UID: \"a3fc6e9a-1447-41e2-8886-e5ea32e3d353\") " pod="openshift-marketplace/redhat-operators-zffmt" Dec 01 08:42:55 crc kubenswrapper[4873]: I1201 08:42:55.994544 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8s74\" (UniqueName: \"kubernetes.io/projected/a3fc6e9a-1447-41e2-8886-e5ea32e3d353-kube-api-access-b8s74\") pod \"redhat-operators-zffmt\" (UID: \"a3fc6e9a-1447-41e2-8886-e5ea32e3d353\") " pod="openshift-marketplace/redhat-operators-zffmt" Dec 01 08:42:56 crc kubenswrapper[4873]: I1201 08:42:56.111187 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zffmt" Dec 01 08:42:56 crc kubenswrapper[4873]: I1201 08:42:56.180424 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tv8pb"] Dec 01 08:42:56 crc kubenswrapper[4873]: I1201 08:42:56.197081 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tv8pb" Dec 01 08:42:56 crc kubenswrapper[4873]: I1201 08:42:56.209929 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tv8pb"] Dec 01 08:42:56 crc kubenswrapper[4873]: I1201 08:42:56.315663 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23f6a774-b182-4eb6-ba94-74ccc28cf3e9-utilities\") pod \"redhat-operators-tv8pb\" (UID: \"23f6a774-b182-4eb6-ba94-74ccc28cf3e9\") " pod="openshift-marketplace/redhat-operators-tv8pb" Dec 01 08:42:56 crc kubenswrapper[4873]: I1201 08:42:56.315767 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23f6a774-b182-4eb6-ba94-74ccc28cf3e9-catalog-content\") pod \"redhat-operators-tv8pb\" (UID: \"23f6a774-b182-4eb6-ba94-74ccc28cf3e9\") " pod="openshift-marketplace/redhat-operators-tv8pb" Dec 01 08:42:56 crc kubenswrapper[4873]: I1201 08:42:56.315810 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gn56g\" (UniqueName: \"kubernetes.io/projected/23f6a774-b182-4eb6-ba94-74ccc28cf3e9-kube-api-access-gn56g\") pod \"redhat-operators-tv8pb\" (UID: \"23f6a774-b182-4eb6-ba94-74ccc28cf3e9\") " pod="openshift-marketplace/redhat-operators-tv8pb" Dec 01 08:42:56 crc kubenswrapper[4873]: I1201 08:42:56.331143 4873 patch_prober.go:28] interesting pod/router-default-5444994796-h9kpk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 08:42:56 crc kubenswrapper[4873]: [-]has-synced failed: reason withheld Dec 01 08:42:56 crc kubenswrapper[4873]: [+]process-running ok Dec 01 08:42:56 crc kubenswrapper[4873]: healthz check failed Dec 01 08:42:56 crc kubenswrapper[4873]: I1201 08:42:56.331218 4873 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h9kpk" podUID="e686de65-3d86-4826-9f0a-4132c81e9ed2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 08:42:56 crc kubenswrapper[4873]: I1201 08:42:56.418793 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23f6a774-b182-4eb6-ba94-74ccc28cf3e9-utilities\") pod \"redhat-operators-tv8pb\" (UID: \"23f6a774-b182-4eb6-ba94-74ccc28cf3e9\") " pod="openshift-marketplace/redhat-operators-tv8pb" Dec 01 08:42:56 crc kubenswrapper[4873]: I1201 08:42:56.419395 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23f6a774-b182-4eb6-ba94-74ccc28cf3e9-catalog-content\") pod \"redhat-operators-tv8pb\" (UID: \"23f6a774-b182-4eb6-ba94-74ccc28cf3e9\") " pod="openshift-marketplace/redhat-operators-tv8pb" Dec 01 08:42:56 crc kubenswrapper[4873]: I1201 08:42:56.419430 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gn56g\" (UniqueName: \"kubernetes.io/projected/23f6a774-b182-4eb6-ba94-74ccc28cf3e9-kube-api-access-gn56g\") pod \"redhat-operators-tv8pb\" (UID: \"23f6a774-b182-4eb6-ba94-74ccc28cf3e9\") " pod="openshift-marketplace/redhat-operators-tv8pb" Dec 01 08:42:56 crc kubenswrapper[4873]: I1201 08:42:56.420422 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23f6a774-b182-4eb6-ba94-74ccc28cf3e9-catalog-content\") pod \"redhat-operators-tv8pb\" (UID: \"23f6a774-b182-4eb6-ba94-74ccc28cf3e9\") " pod="openshift-marketplace/redhat-operators-tv8pb" Dec 01 08:42:56 crc kubenswrapper[4873]: I1201 08:42:56.420654 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23f6a774-b182-4eb6-ba94-74ccc28cf3e9-utilities\") pod \"redhat-operators-tv8pb\" (UID: \"23f6a774-b182-4eb6-ba94-74ccc28cf3e9\") " pod="openshift-marketplace/redhat-operators-tv8pb" Dec 01 08:42:56 crc kubenswrapper[4873]: I1201 08:42:56.474758 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gn56g\" (UniqueName: \"kubernetes.io/projected/23f6a774-b182-4eb6-ba94-74ccc28cf3e9-kube-api-access-gn56g\") pod \"redhat-operators-tv8pb\" (UID: \"23f6a774-b182-4eb6-ba94-74ccc28cf3e9\") " pod="openshift-marketplace/redhat-operators-tv8pb" Dec 01 08:42:56 crc kubenswrapper[4873]: I1201 08:42:56.580638 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tv8pb" Dec 01 08:42:56 crc kubenswrapper[4873]: I1201 08:42:56.597246 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 08:42:56 crc kubenswrapper[4873]: I1201 08:42:56.680330 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b598cb89-a739-44ff-9a29-6e3da327b52a-kube-api-access\") pod \"b598cb89-a739-44ff-9a29-6e3da327b52a\" (UID: \"b598cb89-a739-44ff-9a29-6e3da327b52a\") " Dec 01 08:42:56 crc kubenswrapper[4873]: I1201 08:42:56.681984 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b598cb89-a739-44ff-9a29-6e3da327b52a-kubelet-dir\") pod \"b598cb89-a739-44ff-9a29-6e3da327b52a\" (UID: \"b598cb89-a739-44ff-9a29-6e3da327b52a\") " Dec 01 08:42:56 crc kubenswrapper[4873]: I1201 08:42:56.682348 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b598cb89-a739-44ff-9a29-6e3da327b52a-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "b598cb89-a739-44ff-9a29-6e3da327b52a" (UID: "b598cb89-a739-44ff-9a29-6e3da327b52a"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:42:56 crc kubenswrapper[4873]: I1201 08:42:56.683123 4873 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b598cb89-a739-44ff-9a29-6e3da327b52a-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 01 08:42:56 crc kubenswrapper[4873]: I1201 08:42:56.697419 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zffmt"] Dec 01 08:42:56 crc kubenswrapper[4873]: I1201 08:42:56.727921 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b598cb89-a739-44ff-9a29-6e3da327b52a-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "b598cb89-a739-44ff-9a29-6e3da327b52a" (UID: "b598cb89-a739-44ff-9a29-6e3da327b52a"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:42:56 crc kubenswrapper[4873]: I1201 08:42:56.783814 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b598cb89-a739-44ff-9a29-6e3da327b52a-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 08:42:56 crc kubenswrapper[4873]: I1201 08:42:56.844089 4873 generic.go:334] "Generic (PLEG): container finished" podID="8c34ba9e-07db-478a-8a30-8622f4de4828" containerID="d723dce650a8070bc846093588c025e8ce3dc6d9297fddd792b8d49de06f8769" exitCode=0 Dec 01 08:42:56 crc kubenswrapper[4873]: I1201 08:42:56.844174 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q7s8h" event={"ID":"8c34ba9e-07db-478a-8a30-8622f4de4828","Type":"ContainerDied","Data":"d723dce650a8070bc846093588c025e8ce3dc6d9297fddd792b8d49de06f8769"} Dec 01 08:42:56 crc kubenswrapper[4873]: I1201 08:42:56.865354 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zffmt" event={"ID":"a3fc6e9a-1447-41e2-8886-e5ea32e3d353","Type":"ContainerStarted","Data":"f803672e16cf9ec4069e2f400930ee609bf4f10cae3d61260a212f8ef760cc9b"} Dec 01 08:42:56 crc kubenswrapper[4873]: I1201 08:42:56.890104 4873 generic.go:334] "Generic (PLEG): container finished" podID="73df40c6-5524-496f-92ba-36100c4af8bb" containerID="c7133b53533ce8651130c3291d66c8d102fb954b3e2ccf5164c777a7182e66fb" exitCode=0 Dec 01 08:42:56 crc kubenswrapper[4873]: I1201 08:42:56.890217 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-spckc" event={"ID":"73df40c6-5524-496f-92ba-36100c4af8bb","Type":"ContainerDied","Data":"c7133b53533ce8651130c3291d66c8d102fb954b3e2ccf5164c777a7182e66fb"} Dec 01 08:42:56 crc kubenswrapper[4873]: I1201 08:42:56.894436 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"b598cb89-a739-44ff-9a29-6e3da327b52a","Type":"ContainerDied","Data":"cdf61c7e9faed9b6b580aba2eb576bd092a47e9e427e4de3740dd7c262967b1c"} Dec 01 08:42:56 crc kubenswrapper[4873]: I1201 08:42:56.894465 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cdf61c7e9faed9b6b580aba2eb576bd092a47e9e427e4de3740dd7c262967b1c" Dec 01 08:42:56 crc kubenswrapper[4873]: I1201 08:42:56.894516 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 08:42:56 crc kubenswrapper[4873]: I1201 08:42:56.946728 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"143397c7-7625-40da-941c-0fe2f2ef6e69","Type":"ContainerStarted","Data":"195958208c4da8f85e83bd8d0c34abfad32592928e9d92e14794603393cbdf18"} Dec 01 08:42:56 crc kubenswrapper[4873]: I1201 08:42:56.947218 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"143397c7-7625-40da-941c-0fe2f2ef6e69","Type":"ContainerStarted","Data":"6a6d93ff5bd6a357abf6063fdf8e68475486f8c5c0de7a982773b364d4decb9c"} Dec 01 08:42:56 crc kubenswrapper[4873]: I1201 08:42:56.979390 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=1.979359642 podStartE2EDuration="1.979359642s" podCreationTimestamp="2025-12-01 08:42:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:56.959809822 +0000 UTC m=+152.861918371" watchObservedRunningTime="2025-12-01 08:42:56.979359642 +0000 UTC m=+152.881468171" Dec 01 08:42:57 crc kubenswrapper[4873]: I1201 08:42:57.136047 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tv8pb"] Dec 01 08:42:57 crc kubenswrapper[4873]: W1201 08:42:57.179662 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod23f6a774_b182_4eb6_ba94_74ccc28cf3e9.slice/crio-22fd2abffbf8c8666b3e6e919390a67cac4fa1e4d701458b010bbbe765e4e6b8 WatchSource:0}: Error finding container 22fd2abffbf8c8666b3e6e919390a67cac4fa1e4d701458b010bbbe765e4e6b8: Status 404 returned error can't find the container with id 22fd2abffbf8c8666b3e6e919390a67cac4fa1e4d701458b010bbbe765e4e6b8 Dec 01 08:42:57 crc kubenswrapper[4873]: I1201 08:42:57.330354 4873 patch_prober.go:28] interesting pod/router-default-5444994796-h9kpk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 08:42:57 crc kubenswrapper[4873]: [-]has-synced failed: reason withheld Dec 01 08:42:57 crc kubenswrapper[4873]: [+]process-running ok Dec 01 08:42:57 crc kubenswrapper[4873]: healthz check failed Dec 01 08:42:57 crc kubenswrapper[4873]: I1201 08:42:57.330447 4873 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h9kpk" podUID="e686de65-3d86-4826-9f0a-4132c81e9ed2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 08:42:57 crc kubenswrapper[4873]: I1201 08:42:57.976326 4873 generic.go:334] "Generic (PLEG): container finished" podID="a3fc6e9a-1447-41e2-8886-e5ea32e3d353" containerID="e030f35f012ce934c7de7c1f6faee445f1761fb194016df182d3be267a7a0c88" exitCode=0 Dec 01 08:42:57 crc kubenswrapper[4873]: I1201 08:42:57.976481 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zffmt" event={"ID":"a3fc6e9a-1447-41e2-8886-e5ea32e3d353","Type":"ContainerDied","Data":"e030f35f012ce934c7de7c1f6faee445f1761fb194016df182d3be267a7a0c88"} Dec 01 08:42:57 crc kubenswrapper[4873]: I1201 08:42:57.981106 4873 generic.go:334] "Generic (PLEG): container finished" podID="23f6a774-b182-4eb6-ba94-74ccc28cf3e9" containerID="501b213f2bfe737843050d0aae9436dea62d006452799a47c549cfdb2ad41570" exitCode=0 Dec 01 08:42:57 crc kubenswrapper[4873]: I1201 08:42:57.981904 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tv8pb" event={"ID":"23f6a774-b182-4eb6-ba94-74ccc28cf3e9","Type":"ContainerDied","Data":"501b213f2bfe737843050d0aae9436dea62d006452799a47c549cfdb2ad41570"} Dec 01 08:42:57 crc kubenswrapper[4873]: I1201 08:42:57.981957 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tv8pb" event={"ID":"23f6a774-b182-4eb6-ba94-74ccc28cf3e9","Type":"ContainerStarted","Data":"22fd2abffbf8c8666b3e6e919390a67cac4fa1e4d701458b010bbbe765e4e6b8"} Dec 01 08:42:58 crc kubenswrapper[4873]: I1201 08:42:58.003469 4873 generic.go:334] "Generic (PLEG): container finished" podID="143397c7-7625-40da-941c-0fe2f2ef6e69" containerID="195958208c4da8f85e83bd8d0c34abfad32592928e9d92e14794603393cbdf18" exitCode=0 Dec 01 08:42:58 crc kubenswrapper[4873]: I1201 08:42:58.003518 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"143397c7-7625-40da-941c-0fe2f2ef6e69","Type":"ContainerDied","Data":"195958208c4da8f85e83bd8d0c34abfad32592928e9d92e14794603393cbdf18"} Dec 01 08:42:58 crc kubenswrapper[4873]: I1201 08:42:58.331308 4873 patch_prober.go:28] interesting pod/router-default-5444994796-h9kpk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 08:42:58 crc kubenswrapper[4873]: [-]has-synced failed: reason withheld Dec 01 08:42:58 crc kubenswrapper[4873]: [+]process-running ok Dec 01 08:42:58 crc kubenswrapper[4873]: healthz check failed Dec 01 08:42:58 crc kubenswrapper[4873]: I1201 08:42:58.331432 4873 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h9kpk" podUID="e686de65-3d86-4826-9f0a-4132c81e9ed2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 08:42:59 crc kubenswrapper[4873]: I1201 08:42:59.331270 4873 patch_prober.go:28] interesting pod/router-default-5444994796-h9kpk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 08:42:59 crc kubenswrapper[4873]: [-]has-synced failed: reason withheld Dec 01 08:42:59 crc kubenswrapper[4873]: [+]process-running ok Dec 01 08:42:59 crc kubenswrapper[4873]: healthz check failed Dec 01 08:42:59 crc kubenswrapper[4873]: I1201 08:42:59.331344 4873 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h9kpk" podUID="e686de65-3d86-4826-9f0a-4132c81e9ed2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 08:42:59 crc kubenswrapper[4873]: I1201 08:42:59.459113 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 08:42:59 crc kubenswrapper[4873]: I1201 08:42:59.461975 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-crgj6" Dec 01 08:42:59 crc kubenswrapper[4873]: I1201 08:42:59.684603 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/143397c7-7625-40da-941c-0fe2f2ef6e69-kubelet-dir\") pod \"143397c7-7625-40da-941c-0fe2f2ef6e69\" (UID: \"143397c7-7625-40da-941c-0fe2f2ef6e69\") " Dec 01 08:42:59 crc kubenswrapper[4873]: I1201 08:42:59.684794 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/143397c7-7625-40da-941c-0fe2f2ef6e69-kube-api-access\") pod \"143397c7-7625-40da-941c-0fe2f2ef6e69\" (UID: \"143397c7-7625-40da-941c-0fe2f2ef6e69\") " Dec 01 08:42:59 crc kubenswrapper[4873]: I1201 08:42:59.684797 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/143397c7-7625-40da-941c-0fe2f2ef6e69-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "143397c7-7625-40da-941c-0fe2f2ef6e69" (UID: "143397c7-7625-40da-941c-0fe2f2ef6e69"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:42:59 crc kubenswrapper[4873]: I1201 08:42:59.686488 4873 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/143397c7-7625-40da-941c-0fe2f2ef6e69-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 01 08:42:59 crc kubenswrapper[4873]: I1201 08:42:59.718775 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/143397c7-7625-40da-941c-0fe2f2ef6e69-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "143397c7-7625-40da-941c-0fe2f2ef6e69" (UID: "143397c7-7625-40da-941c-0fe2f2ef6e69"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:42:59 crc kubenswrapper[4873]: I1201 08:42:59.787530 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/143397c7-7625-40da-941c-0fe2f2ef6e69-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:00 crc kubenswrapper[4873]: I1201 08:43:00.047719 4873 patch_prober.go:28] interesting pod/downloads-7954f5f757-76drh container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Dec 01 08:43:00 crc kubenswrapper[4873]: I1201 08:43:00.047813 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-76drh" podUID="dc0adee4-65bf-484b-8f55-3c83022cae82" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Dec 01 08:43:00 crc kubenswrapper[4873]: I1201 08:43:00.047858 4873 patch_prober.go:28] interesting pod/downloads-7954f5f757-76drh container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Dec 01 08:43:00 crc kubenswrapper[4873]: I1201 08:43:00.047953 4873 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-76drh" podUID="dc0adee4-65bf-484b-8f55-3c83022cae82" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Dec 01 08:43:00 crc kubenswrapper[4873]: I1201 08:43:00.108040 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"143397c7-7625-40da-941c-0fe2f2ef6e69","Type":"ContainerDied","Data":"6a6d93ff5bd6a357abf6063fdf8e68475486f8c5c0de7a982773b364d4decb9c"} Dec 01 08:43:00 crc kubenswrapper[4873]: I1201 08:43:00.108089 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6a6d93ff5bd6a357abf6063fdf8e68475486f8c5c0de7a982773b364d4decb9c" Dec 01 08:43:00 crc kubenswrapper[4873]: I1201 08:43:00.108157 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 08:43:00 crc kubenswrapper[4873]: I1201 08:43:00.331795 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-h9kpk" Dec 01 08:43:00 crc kubenswrapper[4873]: I1201 08:43:00.334360 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-h9kpk" Dec 01 08:43:01 crc kubenswrapper[4873]: I1201 08:43:01.031684 4873 patch_prober.go:28] interesting pod/console-f9d7485db-292nl container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.28:8443/health\": dial tcp 10.217.0.28:8443: connect: connection refused" start-of-body= Dec 01 08:43:01 crc kubenswrapper[4873]: I1201 08:43:01.031761 4873 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-292nl" podUID="9491cb1f-acf6-438b-8175-11737d7bd245" containerName="console" probeResult="failure" output="Get \"https://10.217.0.28:8443/health\": dial tcp 10.217.0.28:8443: connect: connection refused" Dec 01 08:43:01 crc kubenswrapper[4873]: I1201 08:43:01.059117 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:43:01 crc kubenswrapper[4873]: I1201 08:43:01.059200 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:43:01 crc kubenswrapper[4873]: I1201 08:43:01.689535 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-7hstm" Dec 01 08:43:06 crc kubenswrapper[4873]: I1201 08:43:06.891435 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/08d02e64-a000-4f11-837e-82ad06d925f5-metrics-certs\") pod \"network-metrics-daemon-9vq2r\" (UID: \"08d02e64-a000-4f11-837e-82ad06d925f5\") " pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:43:06 crc kubenswrapper[4873]: I1201 08:43:06.910388 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/08d02e64-a000-4f11-837e-82ad06d925f5-metrics-certs\") pod \"network-metrics-daemon-9vq2r\" (UID: \"08d02e64-a000-4f11-837e-82ad06d925f5\") " pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:43:06 crc kubenswrapper[4873]: I1201 08:43:06.956485 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9vq2r" Dec 01 08:43:10 crc kubenswrapper[4873]: I1201 08:43:10.057296 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-76drh" Dec 01 08:43:11 crc kubenswrapper[4873]: I1201 08:43:11.139596 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-292nl" Dec 01 08:43:11 crc kubenswrapper[4873]: I1201 08:43:11.150593 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-292nl" Dec 01 08:43:12 crc kubenswrapper[4873]: I1201 08:43:12.745895 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:43:21 crc kubenswrapper[4873]: I1201 08:43:21.146310 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mvczl" Dec 01 08:43:27 crc kubenswrapper[4873]: I1201 08:43:27.665748 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 01 08:43:27 crc kubenswrapper[4873]: E1201 08:43:27.669301 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="143397c7-7625-40da-941c-0fe2f2ef6e69" containerName="pruner" Dec 01 08:43:27 crc kubenswrapper[4873]: I1201 08:43:27.669443 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="143397c7-7625-40da-941c-0fe2f2ef6e69" containerName="pruner" Dec 01 08:43:27 crc kubenswrapper[4873]: E1201 08:43:27.669544 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b598cb89-a739-44ff-9a29-6e3da327b52a" containerName="pruner" Dec 01 08:43:27 crc kubenswrapper[4873]: I1201 08:43:27.669627 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="b598cb89-a739-44ff-9a29-6e3da327b52a" containerName="pruner" Dec 01 08:43:27 crc kubenswrapper[4873]: I1201 08:43:27.669843 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="143397c7-7625-40da-941c-0fe2f2ef6e69" containerName="pruner" Dec 01 08:43:27 crc kubenswrapper[4873]: I1201 08:43:27.669955 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="b598cb89-a739-44ff-9a29-6e3da327b52a" containerName="pruner" Dec 01 08:43:27 crc kubenswrapper[4873]: I1201 08:43:27.670651 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 01 08:43:27 crc kubenswrapper[4873]: I1201 08:43:27.670760 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 08:43:27 crc kubenswrapper[4873]: I1201 08:43:27.672588 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 01 08:43:27 crc kubenswrapper[4873]: I1201 08:43:27.672970 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 01 08:43:27 crc kubenswrapper[4873]: I1201 08:43:27.816500 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/797863da-9202-4834-9e8f-a3bfb45be716-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"797863da-9202-4834-9e8f-a3bfb45be716\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 08:43:27 crc kubenswrapper[4873]: I1201 08:43:27.816576 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/797863da-9202-4834-9e8f-a3bfb45be716-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"797863da-9202-4834-9e8f-a3bfb45be716\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 08:43:27 crc kubenswrapper[4873]: I1201 08:43:27.917782 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/797863da-9202-4834-9e8f-a3bfb45be716-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"797863da-9202-4834-9e8f-a3bfb45be716\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 08:43:27 crc kubenswrapper[4873]: I1201 08:43:27.917888 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/797863da-9202-4834-9e8f-a3bfb45be716-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"797863da-9202-4834-9e8f-a3bfb45be716\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 08:43:27 crc kubenswrapper[4873]: I1201 08:43:27.917990 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/797863da-9202-4834-9e8f-a3bfb45be716-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"797863da-9202-4834-9e8f-a3bfb45be716\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 08:43:27 crc kubenswrapper[4873]: I1201 08:43:27.945312 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/797863da-9202-4834-9e8f-a3bfb45be716-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"797863da-9202-4834-9e8f-a3bfb45be716\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 08:43:27 crc kubenswrapper[4873]: I1201 08:43:27.992324 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 08:43:28 crc kubenswrapper[4873]: E1201 08:43:28.711326 4873 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 01 08:43:28 crc kubenswrapper[4873]: E1201 08:43:28.712291 4873 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vgkbq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-x5tjn_openshift-marketplace(cc58a1d4-9469-4205-abb9-ef87eb162644): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 08:43:28 crc kubenswrapper[4873]: E1201 08:43:28.713940 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-x5tjn" podUID="cc58a1d4-9469-4205-abb9-ef87eb162644" Dec 01 08:43:29 crc kubenswrapper[4873]: I1201 08:43:29.123855 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-9vq2r"] Dec 01 08:43:29 crc kubenswrapper[4873]: W1201 08:43:29.136747 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod08d02e64_a000_4f11_837e_82ad06d925f5.slice/crio-5bd22d0347850823613f54a3058306c9c60e49b870d08fdd73bd475ddd99f783 WatchSource:0}: Error finding container 5bd22d0347850823613f54a3058306c9c60e49b870d08fdd73bd475ddd99f783: Status 404 returned error can't find the container with id 5bd22d0347850823613f54a3058306c9c60e49b870d08fdd73bd475ddd99f783 Dec 01 08:43:29 crc kubenswrapper[4873]: I1201 08:43:29.182456 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 01 08:43:29 crc kubenswrapper[4873]: W1201 08:43:29.216312 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod797863da_9202_4834_9e8f_a3bfb45be716.slice/crio-c89d3547fc9b5ceb77a3eebcfca5702c4085e9e2b36ccde8235d42e72d9fd982 WatchSource:0}: Error finding container c89d3547fc9b5ceb77a3eebcfca5702c4085e9e2b36ccde8235d42e72d9fd982: Status 404 returned error can't find the container with id c89d3547fc9b5ceb77a3eebcfca5702c4085e9e2b36ccde8235d42e72d9fd982 Dec 01 08:43:29 crc kubenswrapper[4873]: I1201 08:43:29.415243 4873 generic.go:334] "Generic (PLEG): container finished" podID="73df40c6-5524-496f-92ba-36100c4af8bb" containerID="6a87de807ce1c794686eccb6fc12010fb7df329698853264ae3cb744cf34effe" exitCode=0 Dec 01 08:43:29 crc kubenswrapper[4873]: I1201 08:43:29.415330 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-spckc" event={"ID":"73df40c6-5524-496f-92ba-36100c4af8bb","Type":"ContainerDied","Data":"6a87de807ce1c794686eccb6fc12010fb7df329698853264ae3cb744cf34effe"} Dec 01 08:43:29 crc kubenswrapper[4873]: I1201 08:43:29.416477 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"797863da-9202-4834-9e8f-a3bfb45be716","Type":"ContainerStarted","Data":"c89d3547fc9b5ceb77a3eebcfca5702c4085e9e2b36ccde8235d42e72d9fd982"} Dec 01 08:43:29 crc kubenswrapper[4873]: I1201 08:43:29.418964 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-9vq2r" event={"ID":"08d02e64-a000-4f11-837e-82ad06d925f5","Type":"ContainerStarted","Data":"5bd22d0347850823613f54a3058306c9c60e49b870d08fdd73bd475ddd99f783"} Dec 01 08:43:29 crc kubenswrapper[4873]: I1201 08:43:29.421844 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zffmt" event={"ID":"a3fc6e9a-1447-41e2-8886-e5ea32e3d353","Type":"ContainerStarted","Data":"895095f340c0d6b6c537639212d41c0b567a44775d21e76ba90c62ffe4bd3ef8"} Dec 01 08:43:29 crc kubenswrapper[4873]: I1201 08:43:29.425772 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jr8nt" event={"ID":"d44d4877-74e9-4fdc-a062-92d031f1636d","Type":"ContainerStarted","Data":"4d52f6e6e3e9a7b6e7d24181236c9333a0e82333e1da5b5d4f9fa17f228e47ff"} Dec 01 08:43:29 crc kubenswrapper[4873]: I1201 08:43:29.429966 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tv8pb" event={"ID":"23f6a774-b182-4eb6-ba94-74ccc28cf3e9","Type":"ContainerStarted","Data":"d4dd8b20aad903a5cab6853de5ad45844621556d1787534f07d001df76db4313"} Dec 01 08:43:29 crc kubenswrapper[4873]: I1201 08:43:29.431949 4873 generic.go:334] "Generic (PLEG): container finished" podID="8c34ba9e-07db-478a-8a30-8622f4de4828" containerID="f5b1d85321e641177cafebbf5da95016c8758fff5aaf26aae663936aa90d5058" exitCode=0 Dec 01 08:43:29 crc kubenswrapper[4873]: I1201 08:43:29.431996 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q7s8h" event={"ID":"8c34ba9e-07db-478a-8a30-8622f4de4828","Type":"ContainerDied","Data":"f5b1d85321e641177cafebbf5da95016c8758fff5aaf26aae663936aa90d5058"} Dec 01 08:43:29 crc kubenswrapper[4873]: I1201 08:43:29.455727 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dm6vl" event={"ID":"7a18563f-c79c-4fed-83fb-5ee7d865e014","Type":"ContainerStarted","Data":"2c7e82be231e5ba969e6f77cb86c791d7aeda2b15bc412839ead5ad6caeddf82"} Dec 01 08:43:29 crc kubenswrapper[4873]: I1201 08:43:29.483208 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8nchr" event={"ID":"1d7debb7-93da-4965-9fb9-fe8626b68183","Type":"ContainerStarted","Data":"768cdc8752ba8ed6fc8382b7579f9852be8fd933d09498887c645de0fe875b49"} Dec 01 08:43:29 crc kubenswrapper[4873]: E1201 08:43:29.489169 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-x5tjn" podUID="cc58a1d4-9469-4205-abb9-ef87eb162644" Dec 01 08:43:30 crc kubenswrapper[4873]: I1201 08:43:30.500843 4873 generic.go:334] "Generic (PLEG): container finished" podID="23f6a774-b182-4eb6-ba94-74ccc28cf3e9" containerID="d4dd8b20aad903a5cab6853de5ad45844621556d1787534f07d001df76db4313" exitCode=0 Dec 01 08:43:30 crc kubenswrapper[4873]: I1201 08:43:30.501041 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tv8pb" event={"ID":"23f6a774-b182-4eb6-ba94-74ccc28cf3e9","Type":"ContainerDied","Data":"d4dd8b20aad903a5cab6853de5ad45844621556d1787534f07d001df76db4313"} Dec 01 08:43:30 crc kubenswrapper[4873]: I1201 08:43:30.515046 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"797863da-9202-4834-9e8f-a3bfb45be716","Type":"ContainerStarted","Data":"e5081aa59e57fcf00316e5acf0ac83e45cd2fb5564c7aaae20782a70ac9211f2"} Dec 01 08:43:30 crc kubenswrapper[4873]: I1201 08:43:30.521362 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-9vq2r" event={"ID":"08d02e64-a000-4f11-837e-82ad06d925f5","Type":"ContainerStarted","Data":"17265aff7fcf032835971b097a0f70f747ea326746123236bb4d597f0ac290ab"} Dec 01 08:43:30 crc kubenswrapper[4873]: I1201 08:43:30.527671 4873 generic.go:334] "Generic (PLEG): container finished" podID="7a18563f-c79c-4fed-83fb-5ee7d865e014" containerID="2c7e82be231e5ba969e6f77cb86c791d7aeda2b15bc412839ead5ad6caeddf82" exitCode=0 Dec 01 08:43:30 crc kubenswrapper[4873]: I1201 08:43:30.527781 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dm6vl" event={"ID":"7a18563f-c79c-4fed-83fb-5ee7d865e014","Type":"ContainerDied","Data":"2c7e82be231e5ba969e6f77cb86c791d7aeda2b15bc412839ead5ad6caeddf82"} Dec 01 08:43:30 crc kubenswrapper[4873]: I1201 08:43:30.539741 4873 generic.go:334] "Generic (PLEG): container finished" podID="1d7debb7-93da-4965-9fb9-fe8626b68183" containerID="768cdc8752ba8ed6fc8382b7579f9852be8fd933d09498887c645de0fe875b49" exitCode=0 Dec 01 08:43:30 crc kubenswrapper[4873]: I1201 08:43:30.539875 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8nchr" event={"ID":"1d7debb7-93da-4965-9fb9-fe8626b68183","Type":"ContainerDied","Data":"768cdc8752ba8ed6fc8382b7579f9852be8fd933d09498887c645de0fe875b49"} Dec 01 08:43:30 crc kubenswrapper[4873]: I1201 08:43:30.547406 4873 generic.go:334] "Generic (PLEG): container finished" podID="a3fc6e9a-1447-41e2-8886-e5ea32e3d353" containerID="895095f340c0d6b6c537639212d41c0b567a44775d21e76ba90c62ffe4bd3ef8" exitCode=0 Dec 01 08:43:30 crc kubenswrapper[4873]: I1201 08:43:30.547501 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zffmt" event={"ID":"a3fc6e9a-1447-41e2-8886-e5ea32e3d353","Type":"ContainerDied","Data":"895095f340c0d6b6c537639212d41c0b567a44775d21e76ba90c62ffe4bd3ef8"} Dec 01 08:43:30 crc kubenswrapper[4873]: I1201 08:43:30.562489 4873 generic.go:334] "Generic (PLEG): container finished" podID="d44d4877-74e9-4fdc-a062-92d031f1636d" containerID="4d52f6e6e3e9a7b6e7d24181236c9333a0e82333e1da5b5d4f9fa17f228e47ff" exitCode=0 Dec 01 08:43:30 crc kubenswrapper[4873]: I1201 08:43:30.562556 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jr8nt" event={"ID":"d44d4877-74e9-4fdc-a062-92d031f1636d","Type":"ContainerDied","Data":"4d52f6e6e3e9a7b6e7d24181236c9333a0e82333e1da5b5d4f9fa17f228e47ff"} Dec 01 08:43:30 crc kubenswrapper[4873]: I1201 08:43:30.582533 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=3.582505819 podStartE2EDuration="3.582505819s" podCreationTimestamp="2025-12-01 08:43:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:43:30.578390509 +0000 UTC m=+186.480499068" watchObservedRunningTime="2025-12-01 08:43:30.582505819 +0000 UTC m=+186.484614358" Dec 01 08:43:30 crc kubenswrapper[4873]: I1201 08:43:30.617733 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:43:31 crc kubenswrapper[4873]: I1201 08:43:31.059481 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:43:31 crc kubenswrapper[4873]: I1201 08:43:31.060061 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:43:31 crc kubenswrapper[4873]: I1201 08:43:31.570615 4873 generic.go:334] "Generic (PLEG): container finished" podID="797863da-9202-4834-9e8f-a3bfb45be716" containerID="e5081aa59e57fcf00316e5acf0ac83e45cd2fb5564c7aaae20782a70ac9211f2" exitCode=0 Dec 01 08:43:31 crc kubenswrapper[4873]: I1201 08:43:31.570699 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"797863da-9202-4834-9e8f-a3bfb45be716","Type":"ContainerDied","Data":"e5081aa59e57fcf00316e5acf0ac83e45cd2fb5564c7aaae20782a70ac9211f2"} Dec 01 08:43:31 crc kubenswrapper[4873]: I1201 08:43:31.573417 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-9vq2r" event={"ID":"08d02e64-a000-4f11-837e-82ad06d925f5","Type":"ContainerStarted","Data":"44581902d0b32e4caa24f398f88b4e0229b0e079f290bf258699dbe82bbb5290"} Dec 01 08:43:31 crc kubenswrapper[4873]: I1201 08:43:31.608707 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-9vq2r" podStartSLOduration=168.608681321 podStartE2EDuration="2m48.608681321s" podCreationTimestamp="2025-12-01 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:43:31.608201821 +0000 UTC m=+187.510310380" watchObservedRunningTime="2025-12-01 08:43:31.608681321 +0000 UTC m=+187.510789860" Dec 01 08:43:32 crc kubenswrapper[4873]: I1201 08:43:32.808696 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 08:43:32 crc kubenswrapper[4873]: I1201 08:43:32.919745 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/797863da-9202-4834-9e8f-a3bfb45be716-kubelet-dir\") pod \"797863da-9202-4834-9e8f-a3bfb45be716\" (UID: \"797863da-9202-4834-9e8f-a3bfb45be716\") " Dec 01 08:43:32 crc kubenswrapper[4873]: I1201 08:43:32.919901 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/797863da-9202-4834-9e8f-a3bfb45be716-kube-api-access\") pod \"797863da-9202-4834-9e8f-a3bfb45be716\" (UID: \"797863da-9202-4834-9e8f-a3bfb45be716\") " Dec 01 08:43:32 crc kubenswrapper[4873]: I1201 08:43:32.919955 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/797863da-9202-4834-9e8f-a3bfb45be716-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "797863da-9202-4834-9e8f-a3bfb45be716" (UID: "797863da-9202-4834-9e8f-a3bfb45be716"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:43:32 crc kubenswrapper[4873]: I1201 08:43:32.920172 4873 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/797863da-9202-4834-9e8f-a3bfb45be716-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:32 crc kubenswrapper[4873]: I1201 08:43:32.927255 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/797863da-9202-4834-9e8f-a3bfb45be716-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "797863da-9202-4834-9e8f-a3bfb45be716" (UID: "797863da-9202-4834-9e8f-a3bfb45be716"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:43:33 crc kubenswrapper[4873]: I1201 08:43:33.021797 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/797863da-9202-4834-9e8f-a3bfb45be716-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:33 crc kubenswrapper[4873]: I1201 08:43:33.585909 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"797863da-9202-4834-9e8f-a3bfb45be716","Type":"ContainerDied","Data":"c89d3547fc9b5ceb77a3eebcfca5702c4085e9e2b36ccde8235d42e72d9fd982"} Dec 01 08:43:33 crc kubenswrapper[4873]: I1201 08:43:33.585951 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 08:43:33 crc kubenswrapper[4873]: I1201 08:43:33.586057 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c89d3547fc9b5ceb77a3eebcfca5702c4085e9e2b36ccde8235d42e72d9fd982" Dec 01 08:43:34 crc kubenswrapper[4873]: I1201 08:43:34.858737 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 01 08:43:34 crc kubenswrapper[4873]: E1201 08:43:34.859447 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="797863da-9202-4834-9e8f-a3bfb45be716" containerName="pruner" Dec 01 08:43:34 crc kubenswrapper[4873]: I1201 08:43:34.859463 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="797863da-9202-4834-9e8f-a3bfb45be716" containerName="pruner" Dec 01 08:43:34 crc kubenswrapper[4873]: I1201 08:43:34.859576 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="797863da-9202-4834-9e8f-a3bfb45be716" containerName="pruner" Dec 01 08:43:34 crc kubenswrapper[4873]: I1201 08:43:34.860061 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 01 08:43:34 crc kubenswrapper[4873]: I1201 08:43:34.862184 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 01 08:43:34 crc kubenswrapper[4873]: I1201 08:43:34.862212 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 01 08:43:34 crc kubenswrapper[4873]: I1201 08:43:34.879183 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 01 08:43:35 crc kubenswrapper[4873]: I1201 08:43:35.049224 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9ecfbd0a-1599-4102-806c-94067f6235bf-kube-api-access\") pod \"installer-9-crc\" (UID: \"9ecfbd0a-1599-4102-806c-94067f6235bf\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 08:43:35 crc kubenswrapper[4873]: I1201 08:43:35.049298 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/9ecfbd0a-1599-4102-806c-94067f6235bf-var-lock\") pod \"installer-9-crc\" (UID: \"9ecfbd0a-1599-4102-806c-94067f6235bf\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 08:43:35 crc kubenswrapper[4873]: I1201 08:43:35.049326 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9ecfbd0a-1599-4102-806c-94067f6235bf-kubelet-dir\") pod \"installer-9-crc\" (UID: \"9ecfbd0a-1599-4102-806c-94067f6235bf\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 08:43:35 crc kubenswrapper[4873]: I1201 08:43:35.150878 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/9ecfbd0a-1599-4102-806c-94067f6235bf-var-lock\") pod \"installer-9-crc\" (UID: \"9ecfbd0a-1599-4102-806c-94067f6235bf\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 08:43:35 crc kubenswrapper[4873]: I1201 08:43:35.150941 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9ecfbd0a-1599-4102-806c-94067f6235bf-kubelet-dir\") pod \"installer-9-crc\" (UID: \"9ecfbd0a-1599-4102-806c-94067f6235bf\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 08:43:35 crc kubenswrapper[4873]: I1201 08:43:35.151103 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9ecfbd0a-1599-4102-806c-94067f6235bf-kube-api-access\") pod \"installer-9-crc\" (UID: \"9ecfbd0a-1599-4102-806c-94067f6235bf\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 08:43:35 crc kubenswrapper[4873]: I1201 08:43:35.151502 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/9ecfbd0a-1599-4102-806c-94067f6235bf-var-lock\") pod \"installer-9-crc\" (UID: \"9ecfbd0a-1599-4102-806c-94067f6235bf\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 08:43:35 crc kubenswrapper[4873]: I1201 08:43:35.151544 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9ecfbd0a-1599-4102-806c-94067f6235bf-kubelet-dir\") pod \"installer-9-crc\" (UID: \"9ecfbd0a-1599-4102-806c-94067f6235bf\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 08:43:35 crc kubenswrapper[4873]: I1201 08:43:35.178178 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9ecfbd0a-1599-4102-806c-94067f6235bf-kube-api-access\") pod \"installer-9-crc\" (UID: \"9ecfbd0a-1599-4102-806c-94067f6235bf\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 08:43:35 crc kubenswrapper[4873]: I1201 08:43:35.188719 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 01 08:43:35 crc kubenswrapper[4873]: I1201 08:43:35.597906 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-spckc" event={"ID":"73df40c6-5524-496f-92ba-36100c4af8bb","Type":"ContainerStarted","Data":"0bb44c93f1cbd0aa1641ba7202566d8d91f2e2a3d3180ac1e2e0897f44fd2fd1"} Dec 01 08:43:35 crc kubenswrapper[4873]: I1201 08:43:35.601344 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q7s8h" event={"ID":"8c34ba9e-07db-478a-8a30-8622f4de4828","Type":"ContainerStarted","Data":"2e9c6ad2eec0a4db9b58cdd760f92eff1e7504423d99e056012d6bf4a6406ce7"} Dec 01 08:43:35 crc kubenswrapper[4873]: I1201 08:43:35.604458 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dm6vl" event={"ID":"7a18563f-c79c-4fed-83fb-5ee7d865e014","Type":"ContainerStarted","Data":"0e4fe44a1413ee54a8747489eee92e455858698931e5b2a4f1dfc41c98e70158"} Dec 01 08:43:35 crc kubenswrapper[4873]: I1201 08:43:35.607655 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8nchr" event={"ID":"1d7debb7-93da-4965-9fb9-fe8626b68183","Type":"ContainerStarted","Data":"bc4a3b17823acc04874cf3c40f60648a4ee6bab2b0450ed0e47cb2feae2d7969"} Dec 01 08:43:35 crc kubenswrapper[4873]: I1201 08:43:35.610042 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zffmt" event={"ID":"a3fc6e9a-1447-41e2-8886-e5ea32e3d353","Type":"ContainerStarted","Data":"f9aaaf1883f656b600c4e791acc271b8a0214606fb549554c4ff0b1a5a3f3977"} Dec 01 08:43:35 crc kubenswrapper[4873]: I1201 08:43:35.611900 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jr8nt" event={"ID":"d44d4877-74e9-4fdc-a062-92d031f1636d","Type":"ContainerStarted","Data":"24ec8dd1d9493b9bda7288ab158dd7d3dfe73a1525a5fb6e20b5770a7a49f8f8"} Dec 01 08:43:35 crc kubenswrapper[4873]: I1201 08:43:35.614116 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tv8pb" event={"ID":"23f6a774-b182-4eb6-ba94-74ccc28cf3e9","Type":"ContainerStarted","Data":"e6ed5e717aa1aa582c6275c7294086de6205f304c00fb69a0701230fb91de80e"} Dec 01 08:43:35 crc kubenswrapper[4873]: I1201 08:43:35.626210 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-spckc" podStartSLOduration=5.828298621 podStartE2EDuration="41.626190351s" podCreationTimestamp="2025-12-01 08:42:54 +0000 UTC" firstStartedPulling="2025-12-01 08:42:56.892731332 +0000 UTC m=+152.794839871" lastFinishedPulling="2025-12-01 08:43:32.690623062 +0000 UTC m=+188.592731601" observedRunningTime="2025-12-01 08:43:35.625095655 +0000 UTC m=+191.527204194" watchObservedRunningTime="2025-12-01 08:43:35.626190351 +0000 UTC m=+191.528298880" Dec 01 08:43:35 crc kubenswrapper[4873]: I1201 08:43:35.642536 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-dm6vl" podStartSLOduration=3.422802383 podStartE2EDuration="43.642514517s" podCreationTimestamp="2025-12-01 08:42:52 +0000 UTC" firstStartedPulling="2025-12-01 08:42:54.766640635 +0000 UTC m=+150.668749194" lastFinishedPulling="2025-12-01 08:43:34.986352789 +0000 UTC m=+190.888461328" observedRunningTime="2025-12-01 08:43:35.641926913 +0000 UTC m=+191.544035472" watchObservedRunningTime="2025-12-01 08:43:35.642514517 +0000 UTC m=+191.544623056" Dec 01 08:43:35 crc kubenswrapper[4873]: I1201 08:43:35.669584 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tv8pb" podStartSLOduration=2.700867165 podStartE2EDuration="39.669559048s" podCreationTimestamp="2025-12-01 08:42:56 +0000 UTC" firstStartedPulling="2025-12-01 08:42:57.991572005 +0000 UTC m=+153.893680544" lastFinishedPulling="2025-12-01 08:43:34.960263888 +0000 UTC m=+190.862372427" observedRunningTime="2025-12-01 08:43:35.668133629 +0000 UTC m=+191.570242168" watchObservedRunningTime="2025-12-01 08:43:35.669559048 +0000 UTC m=+191.571667587" Dec 01 08:43:35 crc kubenswrapper[4873]: I1201 08:43:35.708474 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zffmt" podStartSLOduration=3.514999723 podStartE2EDuration="40.708427218s" podCreationTimestamp="2025-12-01 08:42:55 +0000 UTC" firstStartedPulling="2025-12-01 08:42:57.979987465 +0000 UTC m=+153.882096004" lastFinishedPulling="2025-12-01 08:43:35.17341496 +0000 UTC m=+191.075523499" observedRunningTime="2025-12-01 08:43:35.686133595 +0000 UTC m=+191.588242144" watchObservedRunningTime="2025-12-01 08:43:35.708427218 +0000 UTC m=+191.610535757" Dec 01 08:43:35 crc kubenswrapper[4873]: I1201 08:43:35.709395 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jr8nt" podStartSLOduration=3.483362142 podStartE2EDuration="43.709389058s" podCreationTimestamp="2025-12-01 08:42:52 +0000 UTC" firstStartedPulling="2025-12-01 08:42:54.785055388 +0000 UTC m=+150.687163927" lastFinishedPulling="2025-12-01 08:43:35.011082304 +0000 UTC m=+190.913190843" observedRunningTime="2025-12-01 08:43:35.70556148 +0000 UTC m=+191.607670019" watchObservedRunningTime="2025-12-01 08:43:35.709389058 +0000 UTC m=+191.611497597" Dec 01 08:43:35 crc kubenswrapper[4873]: I1201 08:43:35.729197 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8nchr" podStartSLOduration=3.449272397 podStartE2EDuration="42.729178358s" podCreationTimestamp="2025-12-01 08:42:53 +0000 UTC" firstStartedPulling="2025-12-01 08:42:55.815699355 +0000 UTC m=+151.717807884" lastFinishedPulling="2025-12-01 08:43:35.095605306 +0000 UTC m=+190.997713845" observedRunningTime="2025-12-01 08:43:35.726849572 +0000 UTC m=+191.628958111" watchObservedRunningTime="2025-12-01 08:43:35.729178358 +0000 UTC m=+191.631286897" Dec 01 08:43:35 crc kubenswrapper[4873]: I1201 08:43:35.736427 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 01 08:43:35 crc kubenswrapper[4873]: W1201 08:43:35.752962 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod9ecfbd0a_1599_4102_806c_94067f6235bf.slice/crio-d940d98c8b205baa282afcafa4bcbbd7a230e99430ea7c1a291a4eb0ffd7a470 WatchSource:0}: Error finding container d940d98c8b205baa282afcafa4bcbbd7a230e99430ea7c1a291a4eb0ffd7a470: Status 404 returned error can't find the container with id d940d98c8b205baa282afcafa4bcbbd7a230e99430ea7c1a291a4eb0ffd7a470 Dec 01 08:43:36 crc kubenswrapper[4873]: I1201 08:43:36.112175 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zffmt" Dec 01 08:43:36 crc kubenswrapper[4873]: I1201 08:43:36.112239 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zffmt" Dec 01 08:43:36 crc kubenswrapper[4873]: I1201 08:43:36.581445 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-tv8pb" Dec 01 08:43:36 crc kubenswrapper[4873]: I1201 08:43:36.581972 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tv8pb" Dec 01 08:43:36 crc kubenswrapper[4873]: I1201 08:43:36.622491 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"9ecfbd0a-1599-4102-806c-94067f6235bf","Type":"ContainerStarted","Data":"4d8b3cbf9ad510b5d9e58966d4cddbeb2aba658df386cba296047e50b267a02f"} Dec 01 08:43:36 crc kubenswrapper[4873]: I1201 08:43:36.622572 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"9ecfbd0a-1599-4102-806c-94067f6235bf","Type":"ContainerStarted","Data":"d940d98c8b205baa282afcafa4bcbbd7a230e99430ea7c1a291a4eb0ffd7a470"} Dec 01 08:43:36 crc kubenswrapper[4873]: I1201 08:43:36.639715 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=2.639695897 podStartE2EDuration="2.639695897s" podCreationTimestamp="2025-12-01 08:43:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:43:36.636919372 +0000 UTC m=+192.539027911" watchObservedRunningTime="2025-12-01 08:43:36.639695897 +0000 UTC m=+192.541804436" Dec 01 08:43:36 crc kubenswrapper[4873]: I1201 08:43:36.639921 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-q7s8h" podStartSLOduration=5.178234164 podStartE2EDuration="42.639915226s" podCreationTimestamp="2025-12-01 08:42:54 +0000 UTC" firstStartedPulling="2025-12-01 08:42:56.84731448 +0000 UTC m=+152.749423019" lastFinishedPulling="2025-12-01 08:43:34.308995542 +0000 UTC m=+190.211104081" observedRunningTime="2025-12-01 08:43:35.754593491 +0000 UTC m=+191.656702030" watchObservedRunningTime="2025-12-01 08:43:36.639915226 +0000 UTC m=+192.542023755" Dec 01 08:43:37 crc kubenswrapper[4873]: I1201 08:43:37.185081 4873 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zffmt" podUID="a3fc6e9a-1447-41e2-8886-e5ea32e3d353" containerName="registry-server" probeResult="failure" output=< Dec 01 08:43:37 crc kubenswrapper[4873]: timeout: failed to connect service ":50051" within 1s Dec 01 08:43:37 crc kubenswrapper[4873]: > Dec 01 08:43:37 crc kubenswrapper[4873]: I1201 08:43:37.637103 4873 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-tv8pb" podUID="23f6a774-b182-4eb6-ba94-74ccc28cf3e9" containerName="registry-server" probeResult="failure" output=< Dec 01 08:43:37 crc kubenswrapper[4873]: timeout: failed to connect service ":50051" within 1s Dec 01 08:43:37 crc kubenswrapper[4873]: > Dec 01 08:43:42 crc kubenswrapper[4873]: I1201 08:43:42.988404 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jr8nt" Dec 01 08:43:42 crc kubenswrapper[4873]: I1201 08:43:42.989110 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jr8nt" Dec 01 08:43:43 crc kubenswrapper[4873]: I1201 08:43:43.068945 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jr8nt" Dec 01 08:43:43 crc kubenswrapper[4873]: I1201 08:43:43.208117 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-dm6vl" Dec 01 08:43:43 crc kubenswrapper[4873]: I1201 08:43:43.208612 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-dm6vl" Dec 01 08:43:43 crc kubenswrapper[4873]: I1201 08:43:43.259905 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-dm6vl" Dec 01 08:43:43 crc kubenswrapper[4873]: I1201 08:43:43.525229 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8nchr" Dec 01 08:43:43 crc kubenswrapper[4873]: I1201 08:43:43.525285 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8nchr" Dec 01 08:43:43 crc kubenswrapper[4873]: I1201 08:43:43.563858 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8nchr" Dec 01 08:43:43 crc kubenswrapper[4873]: I1201 08:43:43.705568 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-dm6vl" Dec 01 08:43:43 crc kubenswrapper[4873]: I1201 08:43:43.717458 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jr8nt" Dec 01 08:43:43 crc kubenswrapper[4873]: I1201 08:43:43.719159 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8nchr" Dec 01 08:43:45 crc kubenswrapper[4873]: I1201 08:43:45.051255 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-spckc" Dec 01 08:43:45 crc kubenswrapper[4873]: I1201 08:43:45.051978 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-spckc" Dec 01 08:43:45 crc kubenswrapper[4873]: I1201 08:43:45.114327 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-spckc" Dec 01 08:43:45 crc kubenswrapper[4873]: I1201 08:43:45.413287 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-q7s8h" Dec 01 08:43:45 crc kubenswrapper[4873]: I1201 08:43:45.413555 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-q7s8h" Dec 01 08:43:45 crc kubenswrapper[4873]: I1201 08:43:45.457230 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8nchr"] Dec 01 08:43:45 crc kubenswrapper[4873]: I1201 08:43:45.468235 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-q7s8h" Dec 01 08:43:45 crc kubenswrapper[4873]: I1201 08:43:45.725170 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8nchr" podUID="1d7debb7-93da-4965-9fb9-fe8626b68183" containerName="registry-server" containerID="cri-o://bc4a3b17823acc04874cf3c40f60648a4ee6bab2b0450ed0e47cb2feae2d7969" gracePeriod=2 Dec 01 08:43:45 crc kubenswrapper[4873]: I1201 08:43:45.768709 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-q7s8h" Dec 01 08:43:45 crc kubenswrapper[4873]: I1201 08:43:45.771202 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-spckc" Dec 01 08:43:46 crc kubenswrapper[4873]: I1201 08:43:46.157690 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zffmt" Dec 01 08:43:46 crc kubenswrapper[4873]: I1201 08:43:46.201251 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zffmt" Dec 01 08:43:46 crc kubenswrapper[4873]: I1201 08:43:46.641374 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tv8pb" Dec 01 08:43:46 crc kubenswrapper[4873]: I1201 08:43:46.691454 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tv8pb" Dec 01 08:43:47 crc kubenswrapper[4873]: I1201 08:43:47.864657 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q7s8h"] Dec 01 08:43:48 crc kubenswrapper[4873]: I1201 08:43:48.567550 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-8nchr_1d7debb7-93da-4965-9fb9-fe8626b68183/registry-server/0.log" Dec 01 08:43:48 crc kubenswrapper[4873]: I1201 08:43:48.568384 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8nchr" Dec 01 08:43:48 crc kubenswrapper[4873]: I1201 08:43:48.701960 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d7debb7-93da-4965-9fb9-fe8626b68183-utilities\") pod \"1d7debb7-93da-4965-9fb9-fe8626b68183\" (UID: \"1d7debb7-93da-4965-9fb9-fe8626b68183\") " Dec 01 08:43:48 crc kubenswrapper[4873]: I1201 08:43:48.702151 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d7debb7-93da-4965-9fb9-fe8626b68183-catalog-content\") pod \"1d7debb7-93da-4965-9fb9-fe8626b68183\" (UID: \"1d7debb7-93da-4965-9fb9-fe8626b68183\") " Dec 01 08:43:48 crc kubenswrapper[4873]: I1201 08:43:48.702230 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67mng\" (UniqueName: \"kubernetes.io/projected/1d7debb7-93da-4965-9fb9-fe8626b68183-kube-api-access-67mng\") pod \"1d7debb7-93da-4965-9fb9-fe8626b68183\" (UID: \"1d7debb7-93da-4965-9fb9-fe8626b68183\") " Dec 01 08:43:48 crc kubenswrapper[4873]: I1201 08:43:48.703876 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d7debb7-93da-4965-9fb9-fe8626b68183-utilities" (OuterVolumeSpecName: "utilities") pod "1d7debb7-93da-4965-9fb9-fe8626b68183" (UID: "1d7debb7-93da-4965-9fb9-fe8626b68183"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:43:48 crc kubenswrapper[4873]: I1201 08:43:48.710874 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d7debb7-93da-4965-9fb9-fe8626b68183-kube-api-access-67mng" (OuterVolumeSpecName: "kube-api-access-67mng") pod "1d7debb7-93da-4965-9fb9-fe8626b68183" (UID: "1d7debb7-93da-4965-9fb9-fe8626b68183"). InnerVolumeSpecName "kube-api-access-67mng". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:43:48 crc kubenswrapper[4873]: I1201 08:43:48.750863 4873 generic.go:334] "Generic (PLEG): container finished" podID="cc58a1d4-9469-4205-abb9-ef87eb162644" containerID="16b857d3e3f20e6c0f499142fbfff2f06f7b30bb0ca9c522f261b08598d3b99c" exitCode=0 Dec 01 08:43:48 crc kubenswrapper[4873]: I1201 08:43:48.750944 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x5tjn" event={"ID":"cc58a1d4-9469-4205-abb9-ef87eb162644","Type":"ContainerDied","Data":"16b857d3e3f20e6c0f499142fbfff2f06f7b30bb0ca9c522f261b08598d3b99c"} Dec 01 08:43:48 crc kubenswrapper[4873]: I1201 08:43:48.753392 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-8nchr_1d7debb7-93da-4965-9fb9-fe8626b68183/registry-server/0.log" Dec 01 08:43:48 crc kubenswrapper[4873]: I1201 08:43:48.757547 4873 generic.go:334] "Generic (PLEG): container finished" podID="1d7debb7-93da-4965-9fb9-fe8626b68183" containerID="bc4a3b17823acc04874cf3c40f60648a4ee6bab2b0450ed0e47cb2feae2d7969" exitCode=137 Dec 01 08:43:48 crc kubenswrapper[4873]: I1201 08:43:48.757658 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8nchr" Dec 01 08:43:48 crc kubenswrapper[4873]: I1201 08:43:48.757651 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8nchr" event={"ID":"1d7debb7-93da-4965-9fb9-fe8626b68183","Type":"ContainerDied","Data":"bc4a3b17823acc04874cf3c40f60648a4ee6bab2b0450ed0e47cb2feae2d7969"} Dec 01 08:43:48 crc kubenswrapper[4873]: I1201 08:43:48.758469 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8nchr" event={"ID":"1d7debb7-93da-4965-9fb9-fe8626b68183","Type":"ContainerDied","Data":"ccbccaac54d39aeb561dbebb217e4073d3f172fc70459c35f8baec69f5724f8f"} Dec 01 08:43:48 crc kubenswrapper[4873]: I1201 08:43:48.758504 4873 scope.go:117] "RemoveContainer" containerID="bc4a3b17823acc04874cf3c40f60648a4ee6bab2b0450ed0e47cb2feae2d7969" Dec 01 08:43:48 crc kubenswrapper[4873]: I1201 08:43:48.758547 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-q7s8h" podUID="8c34ba9e-07db-478a-8a30-8622f4de4828" containerName="registry-server" containerID="cri-o://2e9c6ad2eec0a4db9b58cdd760f92eff1e7504423d99e056012d6bf4a6406ce7" gracePeriod=2 Dec 01 08:43:48 crc kubenswrapper[4873]: I1201 08:43:48.775187 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d7debb7-93da-4965-9fb9-fe8626b68183-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d7debb7-93da-4965-9fb9-fe8626b68183" (UID: "1d7debb7-93da-4965-9fb9-fe8626b68183"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:43:48 crc kubenswrapper[4873]: I1201 08:43:48.790310 4873 scope.go:117] "RemoveContainer" containerID="768cdc8752ba8ed6fc8382b7579f9852be8fd933d09498887c645de0fe875b49" Dec 01 08:43:48 crc kubenswrapper[4873]: I1201 08:43:48.804874 4873 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d7debb7-93da-4965-9fb9-fe8626b68183-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:48 crc kubenswrapper[4873]: I1201 08:43:48.804922 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-67mng\" (UniqueName: \"kubernetes.io/projected/1d7debb7-93da-4965-9fb9-fe8626b68183-kube-api-access-67mng\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:48 crc kubenswrapper[4873]: I1201 08:43:48.804938 4873 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d7debb7-93da-4965-9fb9-fe8626b68183-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:48 crc kubenswrapper[4873]: I1201 08:43:48.818441 4873 scope.go:117] "RemoveContainer" containerID="c35383335fd4da5bbd0b644dd42172b196363f22a819a12b5d67d527c2c2331f" Dec 01 08:43:48 crc kubenswrapper[4873]: I1201 08:43:48.900394 4873 scope.go:117] "RemoveContainer" containerID="bc4a3b17823acc04874cf3c40f60648a4ee6bab2b0450ed0e47cb2feae2d7969" Dec 01 08:43:48 crc kubenswrapper[4873]: E1201 08:43:48.901081 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc4a3b17823acc04874cf3c40f60648a4ee6bab2b0450ed0e47cb2feae2d7969\": container with ID starting with bc4a3b17823acc04874cf3c40f60648a4ee6bab2b0450ed0e47cb2feae2d7969 not found: ID does not exist" containerID="bc4a3b17823acc04874cf3c40f60648a4ee6bab2b0450ed0e47cb2feae2d7969" Dec 01 08:43:48 crc kubenswrapper[4873]: I1201 08:43:48.901113 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc4a3b17823acc04874cf3c40f60648a4ee6bab2b0450ed0e47cb2feae2d7969"} err="failed to get container status \"bc4a3b17823acc04874cf3c40f60648a4ee6bab2b0450ed0e47cb2feae2d7969\": rpc error: code = NotFound desc = could not find container \"bc4a3b17823acc04874cf3c40f60648a4ee6bab2b0450ed0e47cb2feae2d7969\": container with ID starting with bc4a3b17823acc04874cf3c40f60648a4ee6bab2b0450ed0e47cb2feae2d7969 not found: ID does not exist" Dec 01 08:43:48 crc kubenswrapper[4873]: I1201 08:43:48.901182 4873 scope.go:117] "RemoveContainer" containerID="768cdc8752ba8ed6fc8382b7579f9852be8fd933d09498887c645de0fe875b49" Dec 01 08:43:48 crc kubenswrapper[4873]: E1201 08:43:48.901582 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"768cdc8752ba8ed6fc8382b7579f9852be8fd933d09498887c645de0fe875b49\": container with ID starting with 768cdc8752ba8ed6fc8382b7579f9852be8fd933d09498887c645de0fe875b49 not found: ID does not exist" containerID="768cdc8752ba8ed6fc8382b7579f9852be8fd933d09498887c645de0fe875b49" Dec 01 08:43:48 crc kubenswrapper[4873]: I1201 08:43:48.901606 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"768cdc8752ba8ed6fc8382b7579f9852be8fd933d09498887c645de0fe875b49"} err="failed to get container status \"768cdc8752ba8ed6fc8382b7579f9852be8fd933d09498887c645de0fe875b49\": rpc error: code = NotFound desc = could not find container \"768cdc8752ba8ed6fc8382b7579f9852be8fd933d09498887c645de0fe875b49\": container with ID starting with 768cdc8752ba8ed6fc8382b7579f9852be8fd933d09498887c645de0fe875b49 not found: ID does not exist" Dec 01 08:43:48 crc kubenswrapper[4873]: I1201 08:43:48.901625 4873 scope.go:117] "RemoveContainer" containerID="c35383335fd4da5bbd0b644dd42172b196363f22a819a12b5d67d527c2c2331f" Dec 01 08:43:48 crc kubenswrapper[4873]: E1201 08:43:48.901934 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c35383335fd4da5bbd0b644dd42172b196363f22a819a12b5d67d527c2c2331f\": container with ID starting with c35383335fd4da5bbd0b644dd42172b196363f22a819a12b5d67d527c2c2331f not found: ID does not exist" containerID="c35383335fd4da5bbd0b644dd42172b196363f22a819a12b5d67d527c2c2331f" Dec 01 08:43:48 crc kubenswrapper[4873]: I1201 08:43:48.901954 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c35383335fd4da5bbd0b644dd42172b196363f22a819a12b5d67d527c2c2331f"} err="failed to get container status \"c35383335fd4da5bbd0b644dd42172b196363f22a819a12b5d67d527c2c2331f\": rpc error: code = NotFound desc = could not find container \"c35383335fd4da5bbd0b644dd42172b196363f22a819a12b5d67d527c2c2331f\": container with ID starting with c35383335fd4da5bbd0b644dd42172b196363f22a819a12b5d67d527c2c2331f not found: ID does not exist" Dec 01 08:43:49 crc kubenswrapper[4873]: I1201 08:43:49.113513 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q7s8h" Dec 01 08:43:49 crc kubenswrapper[4873]: I1201 08:43:49.127436 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8nchr"] Dec 01 08:43:49 crc kubenswrapper[4873]: I1201 08:43:49.130740 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8nchr"] Dec 01 08:43:49 crc kubenswrapper[4873]: I1201 08:43:49.311217 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c34ba9e-07db-478a-8a30-8622f4de4828-utilities\") pod \"8c34ba9e-07db-478a-8a30-8622f4de4828\" (UID: \"8c34ba9e-07db-478a-8a30-8622f4de4828\") " Dec 01 08:43:49 crc kubenswrapper[4873]: I1201 08:43:49.311333 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c34ba9e-07db-478a-8a30-8622f4de4828-catalog-content\") pod \"8c34ba9e-07db-478a-8a30-8622f4de4828\" (UID: \"8c34ba9e-07db-478a-8a30-8622f4de4828\") " Dec 01 08:43:49 crc kubenswrapper[4873]: I1201 08:43:49.311452 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-px8th\" (UniqueName: \"kubernetes.io/projected/8c34ba9e-07db-478a-8a30-8622f4de4828-kube-api-access-px8th\") pod \"8c34ba9e-07db-478a-8a30-8622f4de4828\" (UID: \"8c34ba9e-07db-478a-8a30-8622f4de4828\") " Dec 01 08:43:49 crc kubenswrapper[4873]: I1201 08:43:49.313043 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c34ba9e-07db-478a-8a30-8622f4de4828-utilities" (OuterVolumeSpecName: "utilities") pod "8c34ba9e-07db-478a-8a30-8622f4de4828" (UID: "8c34ba9e-07db-478a-8a30-8622f4de4828"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:43:49 crc kubenswrapper[4873]: I1201 08:43:49.317853 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c34ba9e-07db-478a-8a30-8622f4de4828-kube-api-access-px8th" (OuterVolumeSpecName: "kube-api-access-px8th") pod "8c34ba9e-07db-478a-8a30-8622f4de4828" (UID: "8c34ba9e-07db-478a-8a30-8622f4de4828"). InnerVolumeSpecName "kube-api-access-px8th". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:43:49 crc kubenswrapper[4873]: I1201 08:43:49.344125 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c34ba9e-07db-478a-8a30-8622f4de4828-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8c34ba9e-07db-478a-8a30-8622f4de4828" (UID: "8c34ba9e-07db-478a-8a30-8622f4de4828"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:43:49 crc kubenswrapper[4873]: I1201 08:43:49.414706 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-px8th\" (UniqueName: \"kubernetes.io/projected/8c34ba9e-07db-478a-8a30-8622f4de4828-kube-api-access-px8th\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:49 crc kubenswrapper[4873]: I1201 08:43:49.414751 4873 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c34ba9e-07db-478a-8a30-8622f4de4828-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:49 crc kubenswrapper[4873]: I1201 08:43:49.414766 4873 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c34ba9e-07db-478a-8a30-8622f4de4828-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:49 crc kubenswrapper[4873]: I1201 08:43:49.768347 4873 generic.go:334] "Generic (PLEG): container finished" podID="8c34ba9e-07db-478a-8a30-8622f4de4828" containerID="2e9c6ad2eec0a4db9b58cdd760f92eff1e7504423d99e056012d6bf4a6406ce7" exitCode=0 Dec 01 08:43:49 crc kubenswrapper[4873]: I1201 08:43:49.768418 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q7s8h" event={"ID":"8c34ba9e-07db-478a-8a30-8622f4de4828","Type":"ContainerDied","Data":"2e9c6ad2eec0a4db9b58cdd760f92eff1e7504423d99e056012d6bf4a6406ce7"} Dec 01 08:43:49 crc kubenswrapper[4873]: I1201 08:43:49.768851 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q7s8h" event={"ID":"8c34ba9e-07db-478a-8a30-8622f4de4828","Type":"ContainerDied","Data":"5a9d6a26628fafc169e37a864d580038b77a3e5306de4c57c20c52610e6acb78"} Dec 01 08:43:49 crc kubenswrapper[4873]: I1201 08:43:49.768896 4873 scope.go:117] "RemoveContainer" containerID="2e9c6ad2eec0a4db9b58cdd760f92eff1e7504423d99e056012d6bf4a6406ce7" Dec 01 08:43:49 crc kubenswrapper[4873]: I1201 08:43:49.768457 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q7s8h" Dec 01 08:43:49 crc kubenswrapper[4873]: I1201 08:43:49.773582 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x5tjn" event={"ID":"cc58a1d4-9469-4205-abb9-ef87eb162644","Type":"ContainerStarted","Data":"a13572fc64222eeece1b8d6e1e6e0d689315da0b86309a09a4a1262cfc6d7a9c"} Dec 01 08:43:49 crc kubenswrapper[4873]: I1201 08:43:49.826668 4873 scope.go:117] "RemoveContainer" containerID="f5b1d85321e641177cafebbf5da95016c8758fff5aaf26aae663936aa90d5058" Dec 01 08:43:49 crc kubenswrapper[4873]: I1201 08:43:49.829143 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-x5tjn" podStartSLOduration=3.121377386 podStartE2EDuration="57.829114422s" podCreationTimestamp="2025-12-01 08:42:52 +0000 UTC" firstStartedPulling="2025-12-01 08:42:54.771610411 +0000 UTC m=+150.673718950" lastFinishedPulling="2025-12-01 08:43:49.479347437 +0000 UTC m=+205.381455986" observedRunningTime="2025-12-01 08:43:49.826955249 +0000 UTC m=+205.729063798" watchObservedRunningTime="2025-12-01 08:43:49.829114422 +0000 UTC m=+205.731222971" Dec 01 08:43:49 crc kubenswrapper[4873]: I1201 08:43:49.861414 4873 scope.go:117] "RemoveContainer" containerID="d723dce650a8070bc846093588c025e8ce3dc6d9297fddd792b8d49de06f8769" Dec 01 08:43:49 crc kubenswrapper[4873]: I1201 08:43:49.861497 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q7s8h"] Dec 01 08:43:49 crc kubenswrapper[4873]: I1201 08:43:49.866846 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-q7s8h"] Dec 01 08:43:49 crc kubenswrapper[4873]: I1201 08:43:49.885062 4873 scope.go:117] "RemoveContainer" containerID="2e9c6ad2eec0a4db9b58cdd760f92eff1e7504423d99e056012d6bf4a6406ce7" Dec 01 08:43:49 crc kubenswrapper[4873]: E1201 08:43:49.886655 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e9c6ad2eec0a4db9b58cdd760f92eff1e7504423d99e056012d6bf4a6406ce7\": container with ID starting with 2e9c6ad2eec0a4db9b58cdd760f92eff1e7504423d99e056012d6bf4a6406ce7 not found: ID does not exist" containerID="2e9c6ad2eec0a4db9b58cdd760f92eff1e7504423d99e056012d6bf4a6406ce7" Dec 01 08:43:49 crc kubenswrapper[4873]: I1201 08:43:49.886713 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e9c6ad2eec0a4db9b58cdd760f92eff1e7504423d99e056012d6bf4a6406ce7"} err="failed to get container status \"2e9c6ad2eec0a4db9b58cdd760f92eff1e7504423d99e056012d6bf4a6406ce7\": rpc error: code = NotFound desc = could not find container \"2e9c6ad2eec0a4db9b58cdd760f92eff1e7504423d99e056012d6bf4a6406ce7\": container with ID starting with 2e9c6ad2eec0a4db9b58cdd760f92eff1e7504423d99e056012d6bf4a6406ce7 not found: ID does not exist" Dec 01 08:43:49 crc kubenswrapper[4873]: I1201 08:43:49.886749 4873 scope.go:117] "RemoveContainer" containerID="f5b1d85321e641177cafebbf5da95016c8758fff5aaf26aae663936aa90d5058" Dec 01 08:43:49 crc kubenswrapper[4873]: E1201 08:43:49.887190 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5b1d85321e641177cafebbf5da95016c8758fff5aaf26aae663936aa90d5058\": container with ID starting with f5b1d85321e641177cafebbf5da95016c8758fff5aaf26aae663936aa90d5058 not found: ID does not exist" containerID="f5b1d85321e641177cafebbf5da95016c8758fff5aaf26aae663936aa90d5058" Dec 01 08:43:49 crc kubenswrapper[4873]: I1201 08:43:49.887226 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5b1d85321e641177cafebbf5da95016c8758fff5aaf26aae663936aa90d5058"} err="failed to get container status \"f5b1d85321e641177cafebbf5da95016c8758fff5aaf26aae663936aa90d5058\": rpc error: code = NotFound desc = could not find container \"f5b1d85321e641177cafebbf5da95016c8758fff5aaf26aae663936aa90d5058\": container with ID starting with f5b1d85321e641177cafebbf5da95016c8758fff5aaf26aae663936aa90d5058 not found: ID does not exist" Dec 01 08:43:49 crc kubenswrapper[4873]: I1201 08:43:49.887250 4873 scope.go:117] "RemoveContainer" containerID="d723dce650a8070bc846093588c025e8ce3dc6d9297fddd792b8d49de06f8769" Dec 01 08:43:49 crc kubenswrapper[4873]: E1201 08:43:49.887522 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d723dce650a8070bc846093588c025e8ce3dc6d9297fddd792b8d49de06f8769\": container with ID starting with d723dce650a8070bc846093588c025e8ce3dc6d9297fddd792b8d49de06f8769 not found: ID does not exist" containerID="d723dce650a8070bc846093588c025e8ce3dc6d9297fddd792b8d49de06f8769" Dec 01 08:43:49 crc kubenswrapper[4873]: I1201 08:43:49.887544 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d723dce650a8070bc846093588c025e8ce3dc6d9297fddd792b8d49de06f8769"} err="failed to get container status \"d723dce650a8070bc846093588c025e8ce3dc6d9297fddd792b8d49de06f8769\": rpc error: code = NotFound desc = could not find container \"d723dce650a8070bc846093588c025e8ce3dc6d9297fddd792b8d49de06f8769\": container with ID starting with d723dce650a8070bc846093588c025e8ce3dc6d9297fddd792b8d49de06f8769 not found: ID does not exist" Dec 01 08:43:50 crc kubenswrapper[4873]: I1201 08:43:50.448379 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d7debb7-93da-4965-9fb9-fe8626b68183" path="/var/lib/kubelet/pods/1d7debb7-93da-4965-9fb9-fe8626b68183/volumes" Dec 01 08:43:50 crc kubenswrapper[4873]: I1201 08:43:50.450797 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c34ba9e-07db-478a-8a30-8622f4de4828" path="/var/lib/kubelet/pods/8c34ba9e-07db-478a-8a30-8622f4de4828/volumes" Dec 01 08:43:50 crc kubenswrapper[4873]: I1201 08:43:50.456332 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tv8pb"] Dec 01 08:43:50 crc kubenswrapper[4873]: I1201 08:43:50.456735 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-tv8pb" podUID="23f6a774-b182-4eb6-ba94-74ccc28cf3e9" containerName="registry-server" containerID="cri-o://e6ed5e717aa1aa582c6275c7294086de6205f304c00fb69a0701230fb91de80e" gracePeriod=2 Dec 01 08:43:50 crc kubenswrapper[4873]: I1201 08:43:50.782573 4873 generic.go:334] "Generic (PLEG): container finished" podID="23f6a774-b182-4eb6-ba94-74ccc28cf3e9" containerID="e6ed5e717aa1aa582c6275c7294086de6205f304c00fb69a0701230fb91de80e" exitCode=0 Dec 01 08:43:50 crc kubenswrapper[4873]: I1201 08:43:50.782635 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tv8pb" event={"ID":"23f6a774-b182-4eb6-ba94-74ccc28cf3e9","Type":"ContainerDied","Data":"e6ed5e717aa1aa582c6275c7294086de6205f304c00fb69a0701230fb91de80e"} Dec 01 08:43:50 crc kubenswrapper[4873]: I1201 08:43:50.844494 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tv8pb" Dec 01 08:43:50 crc kubenswrapper[4873]: I1201 08:43:50.888649 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23f6a774-b182-4eb6-ba94-74ccc28cf3e9-utilities\") pod \"23f6a774-b182-4eb6-ba94-74ccc28cf3e9\" (UID: \"23f6a774-b182-4eb6-ba94-74ccc28cf3e9\") " Dec 01 08:43:50 crc kubenswrapper[4873]: I1201 08:43:50.888776 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gn56g\" (UniqueName: \"kubernetes.io/projected/23f6a774-b182-4eb6-ba94-74ccc28cf3e9-kube-api-access-gn56g\") pod \"23f6a774-b182-4eb6-ba94-74ccc28cf3e9\" (UID: \"23f6a774-b182-4eb6-ba94-74ccc28cf3e9\") " Dec 01 08:43:50 crc kubenswrapper[4873]: I1201 08:43:50.888900 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23f6a774-b182-4eb6-ba94-74ccc28cf3e9-catalog-content\") pod \"23f6a774-b182-4eb6-ba94-74ccc28cf3e9\" (UID: \"23f6a774-b182-4eb6-ba94-74ccc28cf3e9\") " Dec 01 08:43:50 crc kubenswrapper[4873]: I1201 08:43:50.893562 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23f6a774-b182-4eb6-ba94-74ccc28cf3e9-utilities" (OuterVolumeSpecName: "utilities") pod "23f6a774-b182-4eb6-ba94-74ccc28cf3e9" (UID: "23f6a774-b182-4eb6-ba94-74ccc28cf3e9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:43:50 crc kubenswrapper[4873]: I1201 08:43:50.899302 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23f6a774-b182-4eb6-ba94-74ccc28cf3e9-kube-api-access-gn56g" (OuterVolumeSpecName: "kube-api-access-gn56g") pod "23f6a774-b182-4eb6-ba94-74ccc28cf3e9" (UID: "23f6a774-b182-4eb6-ba94-74ccc28cf3e9"). InnerVolumeSpecName "kube-api-access-gn56g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:43:50 crc kubenswrapper[4873]: I1201 08:43:50.990641 4873 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23f6a774-b182-4eb6-ba94-74ccc28cf3e9-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:50 crc kubenswrapper[4873]: I1201 08:43:50.990685 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gn56g\" (UniqueName: \"kubernetes.io/projected/23f6a774-b182-4eb6-ba94-74ccc28cf3e9-kube-api-access-gn56g\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:50 crc kubenswrapper[4873]: I1201 08:43:50.999266 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23f6a774-b182-4eb6-ba94-74ccc28cf3e9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "23f6a774-b182-4eb6-ba94-74ccc28cf3e9" (UID: "23f6a774-b182-4eb6-ba94-74ccc28cf3e9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:43:51 crc kubenswrapper[4873]: I1201 08:43:51.091420 4873 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23f6a774-b182-4eb6-ba94-74ccc28cf3e9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:51 crc kubenswrapper[4873]: I1201 08:43:51.794427 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tv8pb" event={"ID":"23f6a774-b182-4eb6-ba94-74ccc28cf3e9","Type":"ContainerDied","Data":"22fd2abffbf8c8666b3e6e919390a67cac4fa1e4d701458b010bbbe765e4e6b8"} Dec 01 08:43:51 crc kubenswrapper[4873]: I1201 08:43:51.794508 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tv8pb" Dec 01 08:43:51 crc kubenswrapper[4873]: I1201 08:43:51.794530 4873 scope.go:117] "RemoveContainer" containerID="e6ed5e717aa1aa582c6275c7294086de6205f304c00fb69a0701230fb91de80e" Dec 01 08:43:51 crc kubenswrapper[4873]: I1201 08:43:51.824449 4873 scope.go:117] "RemoveContainer" containerID="d4dd8b20aad903a5cab6853de5ad45844621556d1787534f07d001df76db4313" Dec 01 08:43:51 crc kubenswrapper[4873]: I1201 08:43:51.848505 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tv8pb"] Dec 01 08:43:51 crc kubenswrapper[4873]: I1201 08:43:51.855388 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-tv8pb"] Dec 01 08:43:51 crc kubenswrapper[4873]: I1201 08:43:51.864446 4873 scope.go:117] "RemoveContainer" containerID="501b213f2bfe737843050d0aae9436dea62d006452799a47c549cfdb2ad41570" Dec 01 08:43:52 crc kubenswrapper[4873]: I1201 08:43:52.448180 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23f6a774-b182-4eb6-ba94-74ccc28cf3e9" path="/var/lib/kubelet/pods/23f6a774-b182-4eb6-ba94-74ccc28cf3e9/volumes" Dec 01 08:43:53 crc kubenswrapper[4873]: I1201 08:43:53.383450 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-x5tjn" Dec 01 08:43:53 crc kubenswrapper[4873]: I1201 08:43:53.383562 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-x5tjn" Dec 01 08:43:53 crc kubenswrapper[4873]: I1201 08:43:53.422394 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-x5tjn" Dec 01 08:44:00 crc kubenswrapper[4873]: I1201 08:44:00.153887 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-sg6bg"] Dec 01 08:44:01 crc kubenswrapper[4873]: I1201 08:44:01.059903 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:44:01 crc kubenswrapper[4873]: I1201 08:44:01.060531 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:44:01 crc kubenswrapper[4873]: I1201 08:44:01.060620 4873 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" Dec 01 08:44:01 crc kubenswrapper[4873]: I1201 08:44:01.061432 4873 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b67d090dbd4e672a6d3e777da244480eaf270518c68f8eaf2292ef7c366b8da4"} pod="openshift-machine-config-operator/machine-config-daemon-scwpp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 08:44:01 crc kubenswrapper[4873]: I1201 08:44:01.061516 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" containerID="cri-o://b67d090dbd4e672a6d3e777da244480eaf270518c68f8eaf2292ef7c366b8da4" gracePeriod=600 Dec 01 08:44:01 crc kubenswrapper[4873]: I1201 08:44:01.874742 4873 generic.go:334] "Generic (PLEG): container finished" podID="fef7b114-0e07-402d-a37b-315c36011f4b" containerID="b67d090dbd4e672a6d3e777da244480eaf270518c68f8eaf2292ef7c366b8da4" exitCode=0 Dec 01 08:44:01 crc kubenswrapper[4873]: I1201 08:44:01.874798 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" event={"ID":"fef7b114-0e07-402d-a37b-315c36011f4b","Type":"ContainerDied","Data":"b67d090dbd4e672a6d3e777da244480eaf270518c68f8eaf2292ef7c366b8da4"} Dec 01 08:44:01 crc kubenswrapper[4873]: I1201 08:44:01.874838 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" event={"ID":"fef7b114-0e07-402d-a37b-315c36011f4b","Type":"ContainerStarted","Data":"9c30061eb3de4736a094430a40ceb3b59f45bd35deebb43c4c6aac312bafc4a8"} Dec 01 08:44:03 crc kubenswrapper[4873]: I1201 08:44:03.430735 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-x5tjn" Dec 01 08:44:03 crc kubenswrapper[4873]: I1201 08:44:03.481782 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-x5tjn"] Dec 01 08:44:03 crc kubenswrapper[4873]: I1201 08:44:03.889561 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-x5tjn" podUID="cc58a1d4-9469-4205-abb9-ef87eb162644" containerName="registry-server" containerID="cri-o://a13572fc64222eeece1b8d6e1e6e0d689315da0b86309a09a4a1262cfc6d7a9c" gracePeriod=2 Dec 01 08:44:04 crc kubenswrapper[4873]: I1201 08:44:04.376295 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x5tjn" Dec 01 08:44:04 crc kubenswrapper[4873]: I1201 08:44:04.561821 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc58a1d4-9469-4205-abb9-ef87eb162644-catalog-content\") pod \"cc58a1d4-9469-4205-abb9-ef87eb162644\" (UID: \"cc58a1d4-9469-4205-abb9-ef87eb162644\") " Dec 01 08:44:04 crc kubenswrapper[4873]: I1201 08:44:04.561959 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vgkbq\" (UniqueName: \"kubernetes.io/projected/cc58a1d4-9469-4205-abb9-ef87eb162644-kube-api-access-vgkbq\") pod \"cc58a1d4-9469-4205-abb9-ef87eb162644\" (UID: \"cc58a1d4-9469-4205-abb9-ef87eb162644\") " Dec 01 08:44:04 crc kubenswrapper[4873]: I1201 08:44:04.562179 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc58a1d4-9469-4205-abb9-ef87eb162644-utilities\") pod \"cc58a1d4-9469-4205-abb9-ef87eb162644\" (UID: \"cc58a1d4-9469-4205-abb9-ef87eb162644\") " Dec 01 08:44:04 crc kubenswrapper[4873]: I1201 08:44:04.563135 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc58a1d4-9469-4205-abb9-ef87eb162644-utilities" (OuterVolumeSpecName: "utilities") pod "cc58a1d4-9469-4205-abb9-ef87eb162644" (UID: "cc58a1d4-9469-4205-abb9-ef87eb162644"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:44:04 crc kubenswrapper[4873]: I1201 08:44:04.572500 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc58a1d4-9469-4205-abb9-ef87eb162644-kube-api-access-vgkbq" (OuterVolumeSpecName: "kube-api-access-vgkbq") pod "cc58a1d4-9469-4205-abb9-ef87eb162644" (UID: "cc58a1d4-9469-4205-abb9-ef87eb162644"). InnerVolumeSpecName "kube-api-access-vgkbq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:44:04 crc kubenswrapper[4873]: I1201 08:44:04.624547 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc58a1d4-9469-4205-abb9-ef87eb162644-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cc58a1d4-9469-4205-abb9-ef87eb162644" (UID: "cc58a1d4-9469-4205-abb9-ef87eb162644"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:44:04 crc kubenswrapper[4873]: I1201 08:44:04.664090 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vgkbq\" (UniqueName: \"kubernetes.io/projected/cc58a1d4-9469-4205-abb9-ef87eb162644-kube-api-access-vgkbq\") on node \"crc\" DevicePath \"\"" Dec 01 08:44:04 crc kubenswrapper[4873]: I1201 08:44:04.664144 4873 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc58a1d4-9469-4205-abb9-ef87eb162644-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:44:04 crc kubenswrapper[4873]: I1201 08:44:04.664166 4873 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc58a1d4-9469-4205-abb9-ef87eb162644-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:44:04 crc kubenswrapper[4873]: I1201 08:44:04.907880 4873 generic.go:334] "Generic (PLEG): container finished" podID="cc58a1d4-9469-4205-abb9-ef87eb162644" containerID="a13572fc64222eeece1b8d6e1e6e0d689315da0b86309a09a4a1262cfc6d7a9c" exitCode=0 Dec 01 08:44:04 crc kubenswrapper[4873]: I1201 08:44:04.907957 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x5tjn" event={"ID":"cc58a1d4-9469-4205-abb9-ef87eb162644","Type":"ContainerDied","Data":"a13572fc64222eeece1b8d6e1e6e0d689315da0b86309a09a4a1262cfc6d7a9c"} Dec 01 08:44:04 crc kubenswrapper[4873]: I1201 08:44:04.907995 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x5tjn" Dec 01 08:44:04 crc kubenswrapper[4873]: I1201 08:44:04.908046 4873 scope.go:117] "RemoveContainer" containerID="a13572fc64222eeece1b8d6e1e6e0d689315da0b86309a09a4a1262cfc6d7a9c" Dec 01 08:44:04 crc kubenswrapper[4873]: I1201 08:44:04.908009 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x5tjn" event={"ID":"cc58a1d4-9469-4205-abb9-ef87eb162644","Type":"ContainerDied","Data":"2516f86433ffd5a73c2b5bb06e7d5d25c158806b10e8b33d50ae55e2eabe2283"} Dec 01 08:44:04 crc kubenswrapper[4873]: I1201 08:44:04.937967 4873 scope.go:117] "RemoveContainer" containerID="16b857d3e3f20e6c0f499142fbfff2f06f7b30bb0ca9c522f261b08598d3b99c" Dec 01 08:44:04 crc kubenswrapper[4873]: I1201 08:44:04.965585 4873 scope.go:117] "RemoveContainer" containerID="b9ca40c29ce38d12a47b10dec6ee4674c41cdeae2a58f13b83d3db551c095444" Dec 01 08:44:04 crc kubenswrapper[4873]: I1201 08:44:04.967777 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-x5tjn"] Dec 01 08:44:04 crc kubenswrapper[4873]: I1201 08:44:04.970218 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-x5tjn"] Dec 01 08:44:04 crc kubenswrapper[4873]: I1201 08:44:04.986635 4873 scope.go:117] "RemoveContainer" containerID="a13572fc64222eeece1b8d6e1e6e0d689315da0b86309a09a4a1262cfc6d7a9c" Dec 01 08:44:04 crc kubenswrapper[4873]: E1201 08:44:04.987286 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a13572fc64222eeece1b8d6e1e6e0d689315da0b86309a09a4a1262cfc6d7a9c\": container with ID starting with a13572fc64222eeece1b8d6e1e6e0d689315da0b86309a09a4a1262cfc6d7a9c not found: ID does not exist" containerID="a13572fc64222eeece1b8d6e1e6e0d689315da0b86309a09a4a1262cfc6d7a9c" Dec 01 08:44:04 crc kubenswrapper[4873]: I1201 08:44:04.987337 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a13572fc64222eeece1b8d6e1e6e0d689315da0b86309a09a4a1262cfc6d7a9c"} err="failed to get container status \"a13572fc64222eeece1b8d6e1e6e0d689315da0b86309a09a4a1262cfc6d7a9c\": rpc error: code = NotFound desc = could not find container \"a13572fc64222eeece1b8d6e1e6e0d689315da0b86309a09a4a1262cfc6d7a9c\": container with ID starting with a13572fc64222eeece1b8d6e1e6e0d689315da0b86309a09a4a1262cfc6d7a9c not found: ID does not exist" Dec 01 08:44:04 crc kubenswrapper[4873]: I1201 08:44:04.987369 4873 scope.go:117] "RemoveContainer" containerID="16b857d3e3f20e6c0f499142fbfff2f06f7b30bb0ca9c522f261b08598d3b99c" Dec 01 08:44:04 crc kubenswrapper[4873]: E1201 08:44:04.987788 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16b857d3e3f20e6c0f499142fbfff2f06f7b30bb0ca9c522f261b08598d3b99c\": container with ID starting with 16b857d3e3f20e6c0f499142fbfff2f06f7b30bb0ca9c522f261b08598d3b99c not found: ID does not exist" containerID="16b857d3e3f20e6c0f499142fbfff2f06f7b30bb0ca9c522f261b08598d3b99c" Dec 01 08:44:04 crc kubenswrapper[4873]: I1201 08:44:04.987825 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16b857d3e3f20e6c0f499142fbfff2f06f7b30bb0ca9c522f261b08598d3b99c"} err="failed to get container status \"16b857d3e3f20e6c0f499142fbfff2f06f7b30bb0ca9c522f261b08598d3b99c\": rpc error: code = NotFound desc = could not find container \"16b857d3e3f20e6c0f499142fbfff2f06f7b30bb0ca9c522f261b08598d3b99c\": container with ID starting with 16b857d3e3f20e6c0f499142fbfff2f06f7b30bb0ca9c522f261b08598d3b99c not found: ID does not exist" Dec 01 08:44:04 crc kubenswrapper[4873]: I1201 08:44:04.987847 4873 scope.go:117] "RemoveContainer" containerID="b9ca40c29ce38d12a47b10dec6ee4674c41cdeae2a58f13b83d3db551c095444" Dec 01 08:44:04 crc kubenswrapper[4873]: E1201 08:44:04.988235 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9ca40c29ce38d12a47b10dec6ee4674c41cdeae2a58f13b83d3db551c095444\": container with ID starting with b9ca40c29ce38d12a47b10dec6ee4674c41cdeae2a58f13b83d3db551c095444 not found: ID does not exist" containerID="b9ca40c29ce38d12a47b10dec6ee4674c41cdeae2a58f13b83d3db551c095444" Dec 01 08:44:04 crc kubenswrapper[4873]: I1201 08:44:04.988333 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9ca40c29ce38d12a47b10dec6ee4674c41cdeae2a58f13b83d3db551c095444"} err="failed to get container status \"b9ca40c29ce38d12a47b10dec6ee4674c41cdeae2a58f13b83d3db551c095444\": rpc error: code = NotFound desc = could not find container \"b9ca40c29ce38d12a47b10dec6ee4674c41cdeae2a58f13b83d3db551c095444\": container with ID starting with b9ca40c29ce38d12a47b10dec6ee4674c41cdeae2a58f13b83d3db551c095444 not found: ID does not exist" Dec 01 08:44:06 crc kubenswrapper[4873]: I1201 08:44:06.446524 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc58a1d4-9469-4205-abb9-ef87eb162644" path="/var/lib/kubelet/pods/cc58a1d4-9469-4205-abb9-ef87eb162644/volumes" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.767291 4873 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 01 08:44:13 crc kubenswrapper[4873]: E1201 08:44:13.768478 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23f6a774-b182-4eb6-ba94-74ccc28cf3e9" containerName="extract-utilities" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.768497 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="23f6a774-b182-4eb6-ba94-74ccc28cf3e9" containerName="extract-utilities" Dec 01 08:44:13 crc kubenswrapper[4873]: E1201 08:44:13.768513 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d7debb7-93da-4965-9fb9-fe8626b68183" containerName="extract-utilities" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.768522 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d7debb7-93da-4965-9fb9-fe8626b68183" containerName="extract-utilities" Dec 01 08:44:13 crc kubenswrapper[4873]: E1201 08:44:13.768539 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d7debb7-93da-4965-9fb9-fe8626b68183" containerName="registry-server" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.768548 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d7debb7-93da-4965-9fb9-fe8626b68183" containerName="registry-server" Dec 01 08:44:13 crc kubenswrapper[4873]: E1201 08:44:13.768562 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d7debb7-93da-4965-9fb9-fe8626b68183" containerName="extract-content" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.768572 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d7debb7-93da-4965-9fb9-fe8626b68183" containerName="extract-content" Dec 01 08:44:13 crc kubenswrapper[4873]: E1201 08:44:13.768584 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c34ba9e-07db-478a-8a30-8622f4de4828" containerName="extract-utilities" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.768592 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c34ba9e-07db-478a-8a30-8622f4de4828" containerName="extract-utilities" Dec 01 08:44:13 crc kubenswrapper[4873]: E1201 08:44:13.768608 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c34ba9e-07db-478a-8a30-8622f4de4828" containerName="extract-content" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.768617 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c34ba9e-07db-478a-8a30-8622f4de4828" containerName="extract-content" Dec 01 08:44:13 crc kubenswrapper[4873]: E1201 08:44:13.768631 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23f6a774-b182-4eb6-ba94-74ccc28cf3e9" containerName="extract-content" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.768639 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="23f6a774-b182-4eb6-ba94-74ccc28cf3e9" containerName="extract-content" Dec 01 08:44:13 crc kubenswrapper[4873]: E1201 08:44:13.768650 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc58a1d4-9469-4205-abb9-ef87eb162644" containerName="registry-server" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.768659 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc58a1d4-9469-4205-abb9-ef87eb162644" containerName="registry-server" Dec 01 08:44:13 crc kubenswrapper[4873]: E1201 08:44:13.768674 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c34ba9e-07db-478a-8a30-8622f4de4828" containerName="registry-server" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.768683 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c34ba9e-07db-478a-8a30-8622f4de4828" containerName="registry-server" Dec 01 08:44:13 crc kubenswrapper[4873]: E1201 08:44:13.768696 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc58a1d4-9469-4205-abb9-ef87eb162644" containerName="extract-content" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.768705 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc58a1d4-9469-4205-abb9-ef87eb162644" containerName="extract-content" Dec 01 08:44:13 crc kubenswrapper[4873]: E1201 08:44:13.768719 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc58a1d4-9469-4205-abb9-ef87eb162644" containerName="extract-utilities" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.768728 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc58a1d4-9469-4205-abb9-ef87eb162644" containerName="extract-utilities" Dec 01 08:44:13 crc kubenswrapper[4873]: E1201 08:44:13.768741 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23f6a774-b182-4eb6-ba94-74ccc28cf3e9" containerName="registry-server" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.768749 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="23f6a774-b182-4eb6-ba94-74ccc28cf3e9" containerName="registry-server" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.768883 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="23f6a774-b182-4eb6-ba94-74ccc28cf3e9" containerName="registry-server" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.768897 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c34ba9e-07db-478a-8a30-8622f4de4828" containerName="registry-server" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.768909 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d7debb7-93da-4965-9fb9-fe8626b68183" containerName="registry-server" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.768922 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc58a1d4-9469-4205-abb9-ef87eb162644" containerName="registry-server" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.769356 4873 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.769537 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.769691 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9" gracePeriod=15 Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.769809 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9" gracePeriod=15 Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.769840 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7" gracePeriod=15 Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.769963 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568" gracePeriod=15 Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.769831 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a" gracePeriod=15 Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.770376 4873 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 01 08:44:13 crc kubenswrapper[4873]: E1201 08:44:13.770888 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.770904 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 01 08:44:13 crc kubenswrapper[4873]: E1201 08:44:13.770917 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.770923 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 01 08:44:13 crc kubenswrapper[4873]: E1201 08:44:13.770933 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.770940 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 08:44:13 crc kubenswrapper[4873]: E1201 08:44:13.770952 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.770959 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 01 08:44:13 crc kubenswrapper[4873]: E1201 08:44:13.770971 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.770980 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 01 08:44:13 crc kubenswrapper[4873]: E1201 08:44:13.770989 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.770995 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 08:44:13 crc kubenswrapper[4873]: E1201 08:44:13.771004 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.771025 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.771145 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.771158 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.771166 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.771173 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.771183 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.771424 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.808672 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.808774 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.808827 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.808862 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.808899 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.808931 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.808995 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.809114 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:44:13 crc kubenswrapper[4873]: E1201 08:44:13.824469 4873 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.195:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.910934 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.911000 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.911050 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.911080 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.911102 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.911123 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.911143 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.911175 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.911262 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.911311 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.911335 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.911361 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.911387 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.911410 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.911437 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.911471 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.968796 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.971309 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.972893 4873 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9" exitCode=0 Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.972948 4873 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568" exitCode=0 Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.972959 4873 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7" exitCode=0 Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.972969 4873 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a" exitCode=2 Dec 01 08:44:13 crc kubenswrapper[4873]: I1201 08:44:13.973047 4873 scope.go:117] "RemoveContainer" containerID="f8a400f96ed3073f12a776072aa91285daadc6aeb71e0c2ff9ebbc4f9dac5519" Dec 01 08:44:14 crc kubenswrapper[4873]: I1201 08:44:14.126498 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:44:14 crc kubenswrapper[4873]: E1201 08:44:14.157087 4873 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.195:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187d0aefd3319515 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-01 08:44:14.155805973 +0000 UTC m=+230.057914512,LastTimestamp:2025-12-01 08:44:14.155805973 +0000 UTC m=+230.057914512,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 01 08:44:14 crc kubenswrapper[4873]: I1201 08:44:14.433119 4873 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 01 08:44:14 crc kubenswrapper[4873]: I1201 08:44:14.982073 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"457fa84c82e055fb574e97e03a4a78e5ed995026a8c5549a445806c28cb03627"} Dec 01 08:44:14 crc kubenswrapper[4873]: I1201 08:44:14.982747 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"ec2add71f833a458ddda9f28458b6105826a0fcaa6ca929ad4d35becdc582421"} Dec 01 08:44:14 crc kubenswrapper[4873]: E1201 08:44:14.984742 4873 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.195:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:44:14 crc kubenswrapper[4873]: I1201 08:44:14.987806 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 01 08:44:14 crc kubenswrapper[4873]: I1201 08:44:14.991081 4873 generic.go:334] "Generic (PLEG): container finished" podID="9ecfbd0a-1599-4102-806c-94067f6235bf" containerID="4d8b3cbf9ad510b5d9e58966d4cddbeb2aba658df386cba296047e50b267a02f" exitCode=0 Dec 01 08:44:14 crc kubenswrapper[4873]: I1201 08:44:14.991152 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"9ecfbd0a-1599-4102-806c-94067f6235bf","Type":"ContainerDied","Data":"4d8b3cbf9ad510b5d9e58966d4cddbeb2aba658df386cba296047e50b267a02f"} Dec 01 08:44:14 crc kubenswrapper[4873]: I1201 08:44:14.992203 4873 status_manager.go:851] "Failed to get status for pod" podUID="9ecfbd0a-1599-4102-806c-94067f6235bf" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 01 08:44:16 crc kubenswrapper[4873]: I1201 08:44:16.167584 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 01 08:44:16 crc kubenswrapper[4873]: I1201 08:44:16.169268 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:44:16 crc kubenswrapper[4873]: I1201 08:44:16.169890 4873 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 01 08:44:16 crc kubenswrapper[4873]: I1201 08:44:16.170276 4873 status_manager.go:851] "Failed to get status for pod" podUID="9ecfbd0a-1599-4102-806c-94067f6235bf" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 01 08:44:16 crc kubenswrapper[4873]: I1201 08:44:16.233765 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 01 08:44:16 crc kubenswrapper[4873]: I1201 08:44:16.234425 4873 status_manager.go:851] "Failed to get status for pod" podUID="9ecfbd0a-1599-4102-806c-94067f6235bf" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 01 08:44:16 crc kubenswrapper[4873]: I1201 08:44:16.234601 4873 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 01 08:44:16 crc kubenswrapper[4873]: I1201 08:44:16.246817 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/9ecfbd0a-1599-4102-806c-94067f6235bf-var-lock\") pod \"9ecfbd0a-1599-4102-806c-94067f6235bf\" (UID: \"9ecfbd0a-1599-4102-806c-94067f6235bf\") " Dec 01 08:44:16 crc kubenswrapper[4873]: I1201 08:44:16.246848 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 01 08:44:16 crc kubenswrapper[4873]: I1201 08:44:16.246907 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 01 08:44:16 crc kubenswrapper[4873]: I1201 08:44:16.246936 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9ecfbd0a-1599-4102-806c-94067f6235bf-kube-api-access\") pod \"9ecfbd0a-1599-4102-806c-94067f6235bf\" (UID: \"9ecfbd0a-1599-4102-806c-94067f6235bf\") " Dec 01 08:44:16 crc kubenswrapper[4873]: I1201 08:44:16.246959 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9ecfbd0a-1599-4102-806c-94067f6235bf-kubelet-dir\") pod \"9ecfbd0a-1599-4102-806c-94067f6235bf\" (UID: \"9ecfbd0a-1599-4102-806c-94067f6235bf\") " Dec 01 08:44:16 crc kubenswrapper[4873]: I1201 08:44:16.246985 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 01 08:44:16 crc kubenswrapper[4873]: I1201 08:44:16.246963 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ecfbd0a-1599-4102-806c-94067f6235bf-var-lock" (OuterVolumeSpecName: "var-lock") pod "9ecfbd0a-1599-4102-806c-94067f6235bf" (UID: "9ecfbd0a-1599-4102-806c-94067f6235bf"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:44:16 crc kubenswrapper[4873]: I1201 08:44:16.247100 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ecfbd0a-1599-4102-806c-94067f6235bf-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "9ecfbd0a-1599-4102-806c-94067f6235bf" (UID: "9ecfbd0a-1599-4102-806c-94067f6235bf"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:44:16 crc kubenswrapper[4873]: I1201 08:44:16.247056 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:44:16 crc kubenswrapper[4873]: I1201 08:44:16.247124 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:44:16 crc kubenswrapper[4873]: I1201 08:44:16.247107 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:44:16 crc kubenswrapper[4873]: I1201 08:44:16.247264 4873 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 01 08:44:16 crc kubenswrapper[4873]: I1201 08:44:16.247291 4873 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9ecfbd0a-1599-4102-806c-94067f6235bf-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 01 08:44:16 crc kubenswrapper[4873]: I1201 08:44:16.247307 4873 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 01 08:44:16 crc kubenswrapper[4873]: I1201 08:44:16.247324 4873 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/9ecfbd0a-1599-4102-806c-94067f6235bf-var-lock\") on node \"crc\" DevicePath \"\"" Dec 01 08:44:16 crc kubenswrapper[4873]: I1201 08:44:16.247339 4873 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 01 08:44:16 crc kubenswrapper[4873]: I1201 08:44:16.255891 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ecfbd0a-1599-4102-806c-94067f6235bf-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "9ecfbd0a-1599-4102-806c-94067f6235bf" (UID: "9ecfbd0a-1599-4102-806c-94067f6235bf"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:44:16 crc kubenswrapper[4873]: I1201 08:44:16.348438 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9ecfbd0a-1599-4102-806c-94067f6235bf-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 08:44:16 crc kubenswrapper[4873]: I1201 08:44:16.443804 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 01 08:44:17 crc kubenswrapper[4873]: I1201 08:44:17.007274 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"9ecfbd0a-1599-4102-806c-94067f6235bf","Type":"ContainerDied","Data":"d940d98c8b205baa282afcafa4bcbbd7a230e99430ea7c1a291a4eb0ffd7a470"} Dec 01 08:44:17 crc kubenswrapper[4873]: I1201 08:44:17.007347 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d940d98c8b205baa282afcafa4bcbbd7a230e99430ea7c1a291a4eb0ffd7a470" Dec 01 08:44:17 crc kubenswrapper[4873]: I1201 08:44:17.007368 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 01 08:44:17 crc kubenswrapper[4873]: I1201 08:44:17.015800 4873 status_manager.go:851] "Failed to get status for pod" podUID="9ecfbd0a-1599-4102-806c-94067f6235bf" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 01 08:44:17 crc kubenswrapper[4873]: I1201 08:44:17.016216 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 01 08:44:17 crc kubenswrapper[4873]: I1201 08:44:17.018238 4873 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9" exitCode=0 Dec 01 08:44:17 crc kubenswrapper[4873]: I1201 08:44:17.018312 4873 scope.go:117] "RemoveContainer" containerID="28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9" Dec 01 08:44:17 crc kubenswrapper[4873]: I1201 08:44:17.018440 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:44:17 crc kubenswrapper[4873]: I1201 08:44:17.019392 4873 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 01 08:44:17 crc kubenswrapper[4873]: I1201 08:44:17.020053 4873 status_manager.go:851] "Failed to get status for pod" podUID="9ecfbd0a-1599-4102-806c-94067f6235bf" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 01 08:44:17 crc kubenswrapper[4873]: I1201 08:44:17.021961 4873 status_manager.go:851] "Failed to get status for pod" podUID="9ecfbd0a-1599-4102-806c-94067f6235bf" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 01 08:44:17 crc kubenswrapper[4873]: I1201 08:44:17.022331 4873 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 01 08:44:17 crc kubenswrapper[4873]: I1201 08:44:17.039553 4873 scope.go:117] "RemoveContainer" containerID="c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568" Dec 01 08:44:17 crc kubenswrapper[4873]: I1201 08:44:17.057517 4873 scope.go:117] "RemoveContainer" containerID="fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7" Dec 01 08:44:17 crc kubenswrapper[4873]: I1201 08:44:17.073206 4873 scope.go:117] "RemoveContainer" containerID="1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a" Dec 01 08:44:17 crc kubenswrapper[4873]: I1201 08:44:17.091491 4873 scope.go:117] "RemoveContainer" containerID="9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9" Dec 01 08:44:17 crc kubenswrapper[4873]: I1201 08:44:17.107443 4873 scope.go:117] "RemoveContainer" containerID="0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb" Dec 01 08:44:17 crc kubenswrapper[4873]: I1201 08:44:17.138368 4873 scope.go:117] "RemoveContainer" containerID="28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9" Dec 01 08:44:17 crc kubenswrapper[4873]: E1201 08:44:17.138944 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9\": container with ID starting with 28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9 not found: ID does not exist" containerID="28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9" Dec 01 08:44:17 crc kubenswrapper[4873]: I1201 08:44:17.138980 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9"} err="failed to get container status \"28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9\": rpc error: code = NotFound desc = could not find container \"28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9\": container with ID starting with 28d40e4329f83b0e6d6d4537f5bf3b6989facd3f428c11ba15f6519efcdd42a9 not found: ID does not exist" Dec 01 08:44:17 crc kubenswrapper[4873]: I1201 08:44:17.139010 4873 scope.go:117] "RemoveContainer" containerID="c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568" Dec 01 08:44:17 crc kubenswrapper[4873]: E1201 08:44:17.140309 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568\": container with ID starting with c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568 not found: ID does not exist" containerID="c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568" Dec 01 08:44:17 crc kubenswrapper[4873]: I1201 08:44:17.140375 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568"} err="failed to get container status \"c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568\": rpc error: code = NotFound desc = could not find container \"c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568\": container with ID starting with c30418d184570e06b082f883e79ff9ba7a9ee846beae314e16ac506408850568 not found: ID does not exist" Dec 01 08:44:17 crc kubenswrapper[4873]: I1201 08:44:17.140421 4873 scope.go:117] "RemoveContainer" containerID="fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7" Dec 01 08:44:17 crc kubenswrapper[4873]: E1201 08:44:17.141042 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7\": container with ID starting with fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7 not found: ID does not exist" containerID="fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7" Dec 01 08:44:17 crc kubenswrapper[4873]: I1201 08:44:17.141091 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7"} err="failed to get container status \"fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7\": rpc error: code = NotFound desc = could not find container \"fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7\": container with ID starting with fe6a4dc26e1fd59edec54016704d768d4b216c4f43376261c445cf0c20fc2cd7 not found: ID does not exist" Dec 01 08:44:17 crc kubenswrapper[4873]: I1201 08:44:17.141130 4873 scope.go:117] "RemoveContainer" containerID="1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a" Dec 01 08:44:17 crc kubenswrapper[4873]: E1201 08:44:17.141614 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a\": container with ID starting with 1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a not found: ID does not exist" containerID="1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a" Dec 01 08:44:17 crc kubenswrapper[4873]: I1201 08:44:17.141642 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a"} err="failed to get container status \"1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a\": rpc error: code = NotFound desc = could not find container \"1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a\": container with ID starting with 1bebc57b379f415e6bbc3fe3ca23a73001c1a7515735df9756e3f6a7ca98293a not found: ID does not exist" Dec 01 08:44:17 crc kubenswrapper[4873]: I1201 08:44:17.141661 4873 scope.go:117] "RemoveContainer" containerID="9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9" Dec 01 08:44:17 crc kubenswrapper[4873]: E1201 08:44:17.142072 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9\": container with ID starting with 9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9 not found: ID does not exist" containerID="9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9" Dec 01 08:44:17 crc kubenswrapper[4873]: I1201 08:44:17.142099 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9"} err="failed to get container status \"9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9\": rpc error: code = NotFound desc = could not find container \"9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9\": container with ID starting with 9a8c14b7a74402b31fad5d01e52676b04a2efbdf0aa4917586d360a762835fc9 not found: ID does not exist" Dec 01 08:44:17 crc kubenswrapper[4873]: I1201 08:44:17.142113 4873 scope.go:117] "RemoveContainer" containerID="0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb" Dec 01 08:44:17 crc kubenswrapper[4873]: E1201 08:44:17.142670 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\": container with ID starting with 0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb not found: ID does not exist" containerID="0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb" Dec 01 08:44:17 crc kubenswrapper[4873]: I1201 08:44:17.142701 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb"} err="failed to get container status \"0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\": rpc error: code = NotFound desc = could not find container \"0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb\": container with ID starting with 0685b917e897079de11009cb5325b788f19437f1f755c4857f081fdff07e9ccb not found: ID does not exist" Dec 01 08:44:19 crc kubenswrapper[4873]: E1201 08:44:19.547193 4873 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.195:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187d0aefd3319515 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-01 08:44:14.155805973 +0000 UTC m=+230.057914512,LastTimestamp:2025-12-01 08:44:14.155805973 +0000 UTC m=+230.057914512,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 01 08:44:20 crc kubenswrapper[4873]: E1201 08:44:20.855456 4873 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 01 08:44:20 crc kubenswrapper[4873]: E1201 08:44:20.856205 4873 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 01 08:44:20 crc kubenswrapper[4873]: E1201 08:44:20.856485 4873 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 01 08:44:20 crc kubenswrapper[4873]: E1201 08:44:20.856692 4873 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 01 08:44:20 crc kubenswrapper[4873]: E1201 08:44:20.856978 4873 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 01 08:44:20 crc kubenswrapper[4873]: I1201 08:44:20.857052 4873 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 01 08:44:20 crc kubenswrapper[4873]: E1201 08:44:20.857376 4873 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.195:6443: connect: connection refused" interval="200ms" Dec 01 08:44:21 crc kubenswrapper[4873]: E1201 08:44:21.058920 4873 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.195:6443: connect: connection refused" interval="400ms" Dec 01 08:44:21 crc kubenswrapper[4873]: E1201 08:44:21.460226 4873 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.195:6443: connect: connection refused" interval="800ms" Dec 01 08:44:22 crc kubenswrapper[4873]: E1201 08:44:22.262293 4873 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.195:6443: connect: connection refused" interval="1.6s" Dec 01 08:44:23 crc kubenswrapper[4873]: E1201 08:44:23.863562 4873 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.195:6443: connect: connection refused" interval="3.2s" Dec 01 08:44:24 crc kubenswrapper[4873]: I1201 08:44:24.433793 4873 status_manager.go:851] "Failed to get status for pod" podUID="9ecfbd0a-1599-4102-806c-94067f6235bf" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 01 08:44:25 crc kubenswrapper[4873]: I1201 08:44:25.203826 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" podUID="2404e1f8-203a-4055-8f66-21d8867e7a92" containerName="oauth-openshift" containerID="cri-o://a3eb595bf52712b15c741f24833ada335abb9c444fd6b09b00e8a9b21a9e3042" gracePeriod=15 Dec 01 08:44:25 crc kubenswrapper[4873]: E1201 08:44:25.517788 4873 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.195:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" volumeName="registry-storage" Dec 01 08:44:25 crc kubenswrapper[4873]: I1201 08:44:25.674623 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:44:25 crc kubenswrapper[4873]: I1201 08:44:25.675598 4873 status_manager.go:851] "Failed to get status for pod" podUID="2404e1f8-203a-4055-8f66-21d8867e7a92" pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-sg6bg\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 01 08:44:25 crc kubenswrapper[4873]: I1201 08:44:25.676250 4873 status_manager.go:851] "Failed to get status for pod" podUID="9ecfbd0a-1599-4102-806c-94067f6235bf" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 01 08:44:25 crc kubenswrapper[4873]: I1201 08:44:25.821984 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-system-service-ca\") pod \"2404e1f8-203a-4055-8f66-21d8867e7a92\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " Dec 01 08:44:25 crc kubenswrapper[4873]: I1201 08:44:25.822694 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-user-template-error\") pod \"2404e1f8-203a-4055-8f66-21d8867e7a92\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " Dec 01 08:44:25 crc kubenswrapper[4873]: I1201 08:44:25.822794 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-user-template-provider-selection\") pod \"2404e1f8-203a-4055-8f66-21d8867e7a92\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " Dec 01 08:44:25 crc kubenswrapper[4873]: I1201 08:44:25.822840 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2404e1f8-203a-4055-8f66-21d8867e7a92-audit-policies\") pod \"2404e1f8-203a-4055-8f66-21d8867e7a92\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " Dec 01 08:44:25 crc kubenswrapper[4873]: I1201 08:44:25.823208 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-system-session\") pod \"2404e1f8-203a-4055-8f66-21d8867e7a92\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " Dec 01 08:44:25 crc kubenswrapper[4873]: I1201 08:44:25.823286 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-system-ocp-branding-template\") pod \"2404e1f8-203a-4055-8f66-21d8867e7a92\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " Dec 01 08:44:25 crc kubenswrapper[4873]: I1201 08:44:25.823365 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-system-router-certs\") pod \"2404e1f8-203a-4055-8f66-21d8867e7a92\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " Dec 01 08:44:25 crc kubenswrapper[4873]: I1201 08:44:25.823430 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4xv9q\" (UniqueName: \"kubernetes.io/projected/2404e1f8-203a-4055-8f66-21d8867e7a92-kube-api-access-4xv9q\") pod \"2404e1f8-203a-4055-8f66-21d8867e7a92\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " Dec 01 08:44:25 crc kubenswrapper[4873]: I1201 08:44:25.823486 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-system-serving-cert\") pod \"2404e1f8-203a-4055-8f66-21d8867e7a92\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " Dec 01 08:44:25 crc kubenswrapper[4873]: I1201 08:44:25.823528 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2404e1f8-203a-4055-8f66-21d8867e7a92-audit-dir\") pod \"2404e1f8-203a-4055-8f66-21d8867e7a92\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " Dec 01 08:44:25 crc kubenswrapper[4873]: I1201 08:44:25.823619 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-system-trusted-ca-bundle\") pod \"2404e1f8-203a-4055-8f66-21d8867e7a92\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " Dec 01 08:44:25 crc kubenswrapper[4873]: I1201 08:44:25.823673 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-system-cliconfig\") pod \"2404e1f8-203a-4055-8f66-21d8867e7a92\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " Dec 01 08:44:25 crc kubenswrapper[4873]: I1201 08:44:25.823723 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-user-template-login\") pod \"2404e1f8-203a-4055-8f66-21d8867e7a92\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " Dec 01 08:44:25 crc kubenswrapper[4873]: I1201 08:44:25.823793 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-user-idp-0-file-data\") pod \"2404e1f8-203a-4055-8f66-21d8867e7a92\" (UID: \"2404e1f8-203a-4055-8f66-21d8867e7a92\") " Dec 01 08:44:25 crc kubenswrapper[4873]: I1201 08:44:25.823803 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2404e1f8-203a-4055-8f66-21d8867e7a92-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "2404e1f8-203a-4055-8f66-21d8867e7a92" (UID: "2404e1f8-203a-4055-8f66-21d8867e7a92"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:44:25 crc kubenswrapper[4873]: I1201 08:44:25.824180 4873 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2404e1f8-203a-4055-8f66-21d8867e7a92-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 01 08:44:25 crc kubenswrapper[4873]: I1201 08:44:25.824245 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2404e1f8-203a-4055-8f66-21d8867e7a92-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "2404e1f8-203a-4055-8f66-21d8867e7a92" (UID: "2404e1f8-203a-4055-8f66-21d8867e7a92"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:44:25 crc kubenswrapper[4873]: I1201 08:44:25.825181 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "2404e1f8-203a-4055-8f66-21d8867e7a92" (UID: "2404e1f8-203a-4055-8f66-21d8867e7a92"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:44:25 crc kubenswrapper[4873]: I1201 08:44:25.825869 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "2404e1f8-203a-4055-8f66-21d8867e7a92" (UID: "2404e1f8-203a-4055-8f66-21d8867e7a92"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:44:25 crc kubenswrapper[4873]: I1201 08:44:25.826145 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "2404e1f8-203a-4055-8f66-21d8867e7a92" (UID: "2404e1f8-203a-4055-8f66-21d8867e7a92"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:44:25 crc kubenswrapper[4873]: I1201 08:44:25.833124 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "2404e1f8-203a-4055-8f66-21d8867e7a92" (UID: "2404e1f8-203a-4055-8f66-21d8867e7a92"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:44:25 crc kubenswrapper[4873]: I1201 08:44:25.833527 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "2404e1f8-203a-4055-8f66-21d8867e7a92" (UID: "2404e1f8-203a-4055-8f66-21d8867e7a92"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:44:25 crc kubenswrapper[4873]: I1201 08:44:25.834960 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "2404e1f8-203a-4055-8f66-21d8867e7a92" (UID: "2404e1f8-203a-4055-8f66-21d8867e7a92"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:44:25 crc kubenswrapper[4873]: I1201 08:44:25.835688 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2404e1f8-203a-4055-8f66-21d8867e7a92-kube-api-access-4xv9q" (OuterVolumeSpecName: "kube-api-access-4xv9q") pod "2404e1f8-203a-4055-8f66-21d8867e7a92" (UID: "2404e1f8-203a-4055-8f66-21d8867e7a92"). InnerVolumeSpecName "kube-api-access-4xv9q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:44:25 crc kubenswrapper[4873]: I1201 08:44:25.835837 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "2404e1f8-203a-4055-8f66-21d8867e7a92" (UID: "2404e1f8-203a-4055-8f66-21d8867e7a92"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:44:25 crc kubenswrapper[4873]: I1201 08:44:25.836279 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "2404e1f8-203a-4055-8f66-21d8867e7a92" (UID: "2404e1f8-203a-4055-8f66-21d8867e7a92"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:44:25 crc kubenswrapper[4873]: I1201 08:44:25.842218 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "2404e1f8-203a-4055-8f66-21d8867e7a92" (UID: "2404e1f8-203a-4055-8f66-21d8867e7a92"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:44:25 crc kubenswrapper[4873]: I1201 08:44:25.843116 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "2404e1f8-203a-4055-8f66-21d8867e7a92" (UID: "2404e1f8-203a-4055-8f66-21d8867e7a92"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:44:25 crc kubenswrapper[4873]: I1201 08:44:25.843532 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "2404e1f8-203a-4055-8f66-21d8867e7a92" (UID: "2404e1f8-203a-4055-8f66-21d8867e7a92"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:44:25 crc kubenswrapper[4873]: I1201 08:44:25.925662 4873 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:44:25 crc kubenswrapper[4873]: I1201 08:44:25.925741 4873 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 01 08:44:25 crc kubenswrapper[4873]: I1201 08:44:25.925766 4873 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 01 08:44:25 crc kubenswrapper[4873]: I1201 08:44:25.925791 4873 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2404e1f8-203a-4055-8f66-21d8867e7a92-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 01 08:44:25 crc kubenswrapper[4873]: I1201 08:44:25.925810 4873 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 01 08:44:25 crc kubenswrapper[4873]: I1201 08:44:25.925835 4873 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 01 08:44:25 crc kubenswrapper[4873]: I1201 08:44:25.925861 4873 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 01 08:44:25 crc kubenswrapper[4873]: I1201 08:44:25.925890 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4xv9q\" (UniqueName: \"kubernetes.io/projected/2404e1f8-203a-4055-8f66-21d8867e7a92-kube-api-access-4xv9q\") on node \"crc\" DevicePath \"\"" Dec 01 08:44:25 crc kubenswrapper[4873]: I1201 08:44:25.925914 4873 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:44:25 crc kubenswrapper[4873]: I1201 08:44:25.925933 4873 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:44:25 crc kubenswrapper[4873]: I1201 08:44:25.925951 4873 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 01 08:44:25 crc kubenswrapper[4873]: I1201 08:44:25.925969 4873 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 01 08:44:25 crc kubenswrapper[4873]: I1201 08:44:25.925990 4873 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2404e1f8-203a-4055-8f66-21d8867e7a92-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:44:26 crc kubenswrapper[4873]: I1201 08:44:26.092169 4873 generic.go:334] "Generic (PLEG): container finished" podID="2404e1f8-203a-4055-8f66-21d8867e7a92" containerID="a3eb595bf52712b15c741f24833ada335abb9c444fd6b09b00e8a9b21a9e3042" exitCode=0 Dec 01 08:44:26 crc kubenswrapper[4873]: I1201 08:44:26.092253 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" event={"ID":"2404e1f8-203a-4055-8f66-21d8867e7a92","Type":"ContainerDied","Data":"a3eb595bf52712b15c741f24833ada335abb9c444fd6b09b00e8a9b21a9e3042"} Dec 01 08:44:26 crc kubenswrapper[4873]: I1201 08:44:26.092278 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" Dec 01 08:44:26 crc kubenswrapper[4873]: I1201 08:44:26.092319 4873 scope.go:117] "RemoveContainer" containerID="a3eb595bf52712b15c741f24833ada335abb9c444fd6b09b00e8a9b21a9e3042" Dec 01 08:44:26 crc kubenswrapper[4873]: I1201 08:44:26.092298 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" event={"ID":"2404e1f8-203a-4055-8f66-21d8867e7a92","Type":"ContainerDied","Data":"1ba6bf355d3416ca5e8e0786258527e32267090fd584943d2badfd06124aff1f"} Dec 01 08:44:26 crc kubenswrapper[4873]: I1201 08:44:26.093885 4873 status_manager.go:851] "Failed to get status for pod" podUID="2404e1f8-203a-4055-8f66-21d8867e7a92" pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-sg6bg\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 01 08:44:26 crc kubenswrapper[4873]: I1201 08:44:26.094423 4873 status_manager.go:851] "Failed to get status for pod" podUID="9ecfbd0a-1599-4102-806c-94067f6235bf" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 01 08:44:26 crc kubenswrapper[4873]: I1201 08:44:26.123625 4873 status_manager.go:851] "Failed to get status for pod" podUID="2404e1f8-203a-4055-8f66-21d8867e7a92" pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-sg6bg\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 01 08:44:26 crc kubenswrapper[4873]: I1201 08:44:26.124543 4873 status_manager.go:851] "Failed to get status for pod" podUID="9ecfbd0a-1599-4102-806c-94067f6235bf" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 01 08:44:26 crc kubenswrapper[4873]: I1201 08:44:26.132689 4873 scope.go:117] "RemoveContainer" containerID="a3eb595bf52712b15c741f24833ada335abb9c444fd6b09b00e8a9b21a9e3042" Dec 01 08:44:26 crc kubenswrapper[4873]: E1201 08:44:26.133741 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3eb595bf52712b15c741f24833ada335abb9c444fd6b09b00e8a9b21a9e3042\": container with ID starting with a3eb595bf52712b15c741f24833ada335abb9c444fd6b09b00e8a9b21a9e3042 not found: ID does not exist" containerID="a3eb595bf52712b15c741f24833ada335abb9c444fd6b09b00e8a9b21a9e3042" Dec 01 08:44:26 crc kubenswrapper[4873]: I1201 08:44:26.133806 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3eb595bf52712b15c741f24833ada335abb9c444fd6b09b00e8a9b21a9e3042"} err="failed to get container status \"a3eb595bf52712b15c741f24833ada335abb9c444fd6b09b00e8a9b21a9e3042\": rpc error: code = NotFound desc = could not find container \"a3eb595bf52712b15c741f24833ada335abb9c444fd6b09b00e8a9b21a9e3042\": container with ID starting with a3eb595bf52712b15c741f24833ada335abb9c444fd6b09b00e8a9b21a9e3042 not found: ID does not exist" Dec 01 08:44:27 crc kubenswrapper[4873]: E1201 08:44:27.066611 4873 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.195:6443: connect: connection refused" interval="6.4s" Dec 01 08:44:27 crc kubenswrapper[4873]: I1201 08:44:27.429159 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:44:27 crc kubenswrapper[4873]: I1201 08:44:27.430246 4873 status_manager.go:851] "Failed to get status for pod" podUID="2404e1f8-203a-4055-8f66-21d8867e7a92" pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-sg6bg\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 01 08:44:27 crc kubenswrapper[4873]: I1201 08:44:27.430942 4873 status_manager.go:851] "Failed to get status for pod" podUID="9ecfbd0a-1599-4102-806c-94067f6235bf" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 01 08:44:27 crc kubenswrapper[4873]: I1201 08:44:27.454876 4873 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f1c8f149-f188-4759-b2dc-9dc15be1bb13" Dec 01 08:44:27 crc kubenswrapper[4873]: I1201 08:44:27.454938 4873 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f1c8f149-f188-4759-b2dc-9dc15be1bb13" Dec 01 08:44:27 crc kubenswrapper[4873]: E1201 08:44:27.455704 4873 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:44:27 crc kubenswrapper[4873]: I1201 08:44:27.456262 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:44:27 crc kubenswrapper[4873]: W1201 08:44:27.488900 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-e4d8fd9ced003ed51cde035390e1676adba68201fef4e478ae3dd0ca1c2e881e WatchSource:0}: Error finding container e4d8fd9ced003ed51cde035390e1676adba68201fef4e478ae3dd0ca1c2e881e: Status 404 returned error can't find the container with id e4d8fd9ced003ed51cde035390e1676adba68201fef4e478ae3dd0ca1c2e881e Dec 01 08:44:28 crc kubenswrapper[4873]: I1201 08:44:28.109666 4873 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="134e5db9f325b5f6c4a5f66addcb9bbe0ef887b4443e05e95ea654051715b6c1" exitCode=0 Dec 01 08:44:28 crc kubenswrapper[4873]: I1201 08:44:28.109846 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"134e5db9f325b5f6c4a5f66addcb9bbe0ef887b4443e05e95ea654051715b6c1"} Dec 01 08:44:28 crc kubenswrapper[4873]: I1201 08:44:28.110300 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"e4d8fd9ced003ed51cde035390e1676adba68201fef4e478ae3dd0ca1c2e881e"} Dec 01 08:44:28 crc kubenswrapper[4873]: I1201 08:44:28.110852 4873 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f1c8f149-f188-4759-b2dc-9dc15be1bb13" Dec 01 08:44:28 crc kubenswrapper[4873]: I1201 08:44:28.110880 4873 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f1c8f149-f188-4759-b2dc-9dc15be1bb13" Dec 01 08:44:28 crc kubenswrapper[4873]: I1201 08:44:28.111869 4873 status_manager.go:851] "Failed to get status for pod" podUID="9ecfbd0a-1599-4102-806c-94067f6235bf" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 01 08:44:28 crc kubenswrapper[4873]: E1201 08:44:28.112451 4873 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:44:28 crc kubenswrapper[4873]: I1201 08:44:28.112506 4873 status_manager.go:851] "Failed to get status for pod" podUID="2404e1f8-203a-4055-8f66-21d8867e7a92" pod="openshift-authentication/oauth-openshift-558db77b4-sg6bg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-sg6bg\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 01 08:44:29 crc kubenswrapper[4873]: I1201 08:44:29.120238 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 01 08:44:29 crc kubenswrapper[4873]: I1201 08:44:29.120319 4873 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="c3ba3cfd8a611c687414d7ecd6b02b423b351d7ac2a4d041085d4ca3bb2b0709" exitCode=1 Dec 01 08:44:29 crc kubenswrapper[4873]: I1201 08:44:29.120407 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"c3ba3cfd8a611c687414d7ecd6b02b423b351d7ac2a4d041085d4ca3bb2b0709"} Dec 01 08:44:29 crc kubenswrapper[4873]: I1201 08:44:29.121176 4873 scope.go:117] "RemoveContainer" containerID="c3ba3cfd8a611c687414d7ecd6b02b423b351d7ac2a4d041085d4ca3bb2b0709" Dec 01 08:44:29 crc kubenswrapper[4873]: I1201 08:44:29.131444 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"d36c5fd206687f340e9cb3400f9b6a1e7ea364dff1a0be2e2117bb55b789a329"} Dec 01 08:44:29 crc kubenswrapper[4873]: I1201 08:44:29.131492 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"33140a63fa1bc974908b6bf7c2c4248ed3b5de9c93355a4c698ed46c0beb30a3"} Dec 01 08:44:29 crc kubenswrapper[4873]: I1201 08:44:29.131503 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"dc643cc9d88ddae30508d43dffaf02189c7a5aa65524f8a41331b84025b5a48f"} Dec 01 08:44:30 crc kubenswrapper[4873]: I1201 08:44:30.142385 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 01 08:44:30 crc kubenswrapper[4873]: I1201 08:44:30.142907 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"7954b407d12e28f1ad8bc383ec0d8523bd603bd0074fdc1069ff2e140587cee1"} Dec 01 08:44:30 crc kubenswrapper[4873]: I1201 08:44:30.147043 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"5667f429a2121130e63a66c5948c3b109108e1d49efc4816c9f5a0646579f48c"} Dec 01 08:44:30 crc kubenswrapper[4873]: I1201 08:44:30.147084 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"ea0cfbaa3ef7a928dc710b2a8d14f89ffc548b19e2ba2a01819c69e92d3cd65e"} Dec 01 08:44:30 crc kubenswrapper[4873]: I1201 08:44:30.147306 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:44:30 crc kubenswrapper[4873]: I1201 08:44:30.147501 4873 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f1c8f149-f188-4759-b2dc-9dc15be1bb13" Dec 01 08:44:30 crc kubenswrapper[4873]: I1201 08:44:30.147552 4873 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f1c8f149-f188-4759-b2dc-9dc15be1bb13" Dec 01 08:44:32 crc kubenswrapper[4873]: I1201 08:44:32.456727 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:44:32 crc kubenswrapper[4873]: I1201 08:44:32.457214 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:44:32 crc kubenswrapper[4873]: I1201 08:44:32.466324 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:44:34 crc kubenswrapper[4873]: I1201 08:44:34.276064 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:44:35 crc kubenswrapper[4873]: I1201 08:44:35.158450 4873 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:44:35 crc kubenswrapper[4873]: I1201 08:44:35.201869 4873 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f1c8f149-f188-4759-b2dc-9dc15be1bb13" Dec 01 08:44:35 crc kubenswrapper[4873]: I1201 08:44:35.201936 4873 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f1c8f149-f188-4759-b2dc-9dc15be1bb13" Dec 01 08:44:35 crc kubenswrapper[4873]: I1201 08:44:35.207604 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:44:35 crc kubenswrapper[4873]: I1201 08:44:35.210999 4873 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="b5db64de-1f8a-47e1-b07b-848a028e5c24" Dec 01 08:44:36 crc kubenswrapper[4873]: I1201 08:44:36.207327 4873 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f1c8f149-f188-4759-b2dc-9dc15be1bb13" Dec 01 08:44:36 crc kubenswrapper[4873]: I1201 08:44:36.207813 4873 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f1c8f149-f188-4759-b2dc-9dc15be1bb13" Dec 01 08:44:38 crc kubenswrapper[4873]: I1201 08:44:38.442355 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:44:38 crc kubenswrapper[4873]: I1201 08:44:38.445980 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:44:39 crc kubenswrapper[4873]: I1201 08:44:39.231928 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:44:44 crc kubenswrapper[4873]: I1201 08:44:44.457540 4873 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="b5db64de-1f8a-47e1-b07b-848a028e5c24" Dec 01 08:44:45 crc kubenswrapper[4873]: I1201 08:44:45.523326 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 01 08:44:45 crc kubenswrapper[4873]: I1201 08:44:45.872723 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 01 08:44:46 crc kubenswrapper[4873]: I1201 08:44:46.576265 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 01 08:44:46 crc kubenswrapper[4873]: I1201 08:44:46.721843 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 01 08:44:46 crc kubenswrapper[4873]: I1201 08:44:46.762328 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 01 08:44:46 crc kubenswrapper[4873]: I1201 08:44:46.826823 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 01 08:44:46 crc kubenswrapper[4873]: I1201 08:44:46.856344 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 01 08:44:46 crc kubenswrapper[4873]: I1201 08:44:46.972464 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 01 08:44:47 crc kubenswrapper[4873]: I1201 08:44:47.191096 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 01 08:44:47 crc kubenswrapper[4873]: I1201 08:44:47.423996 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 01 08:44:47 crc kubenswrapper[4873]: I1201 08:44:47.759598 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 01 08:44:47 crc kubenswrapper[4873]: I1201 08:44:47.775866 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 01 08:44:47 crc kubenswrapper[4873]: I1201 08:44:47.786169 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 01 08:44:47 crc kubenswrapper[4873]: I1201 08:44:47.850846 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 01 08:44:47 crc kubenswrapper[4873]: I1201 08:44:47.927159 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 01 08:44:47 crc kubenswrapper[4873]: I1201 08:44:47.953687 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 01 08:44:47 crc kubenswrapper[4873]: I1201 08:44:47.998285 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 01 08:44:48 crc kubenswrapper[4873]: I1201 08:44:48.216554 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 01 08:44:48 crc kubenswrapper[4873]: I1201 08:44:48.483885 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 01 08:44:48 crc kubenswrapper[4873]: I1201 08:44:48.522997 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 01 08:44:48 crc kubenswrapper[4873]: I1201 08:44:48.631130 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 01 08:44:48 crc kubenswrapper[4873]: I1201 08:44:48.634481 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 01 08:44:48 crc kubenswrapper[4873]: I1201 08:44:48.688746 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 01 08:44:48 crc kubenswrapper[4873]: I1201 08:44:48.865205 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 01 08:44:48 crc kubenswrapper[4873]: I1201 08:44:48.867916 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 01 08:44:48 crc kubenswrapper[4873]: I1201 08:44:48.886874 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 01 08:44:48 crc kubenswrapper[4873]: I1201 08:44:48.926666 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 01 08:44:48 crc kubenswrapper[4873]: I1201 08:44:48.934252 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 01 08:44:48 crc kubenswrapper[4873]: I1201 08:44:48.982622 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 01 08:44:49 crc kubenswrapper[4873]: I1201 08:44:49.034178 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 01 08:44:49 crc kubenswrapper[4873]: I1201 08:44:49.039713 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 01 08:44:49 crc kubenswrapper[4873]: I1201 08:44:49.065167 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 01 08:44:49 crc kubenswrapper[4873]: I1201 08:44:49.213331 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 01 08:44:49 crc kubenswrapper[4873]: I1201 08:44:49.222412 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 01 08:44:49 crc kubenswrapper[4873]: I1201 08:44:49.246328 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 01 08:44:49 crc kubenswrapper[4873]: I1201 08:44:49.368913 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 01 08:44:49 crc kubenswrapper[4873]: I1201 08:44:49.399486 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 01 08:44:49 crc kubenswrapper[4873]: I1201 08:44:49.416149 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 01 08:44:49 crc kubenswrapper[4873]: I1201 08:44:49.435292 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 01 08:44:49 crc kubenswrapper[4873]: I1201 08:44:49.549090 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 01 08:44:49 crc kubenswrapper[4873]: I1201 08:44:49.595577 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 01 08:44:49 crc kubenswrapper[4873]: I1201 08:44:49.663324 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 01 08:44:49 crc kubenswrapper[4873]: I1201 08:44:49.823046 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 01 08:44:49 crc kubenswrapper[4873]: I1201 08:44:49.846003 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 01 08:44:49 crc kubenswrapper[4873]: I1201 08:44:49.889268 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 01 08:44:49 crc kubenswrapper[4873]: I1201 08:44:49.910346 4873 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 01 08:44:49 crc kubenswrapper[4873]: I1201 08:44:49.972685 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 01 08:44:50 crc kubenswrapper[4873]: I1201 08:44:50.015824 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 01 08:44:50 crc kubenswrapper[4873]: I1201 08:44:50.058183 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 01 08:44:50 crc kubenswrapper[4873]: I1201 08:44:50.112647 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 01 08:44:50 crc kubenswrapper[4873]: I1201 08:44:50.138010 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 01 08:44:50 crc kubenswrapper[4873]: I1201 08:44:50.234173 4873 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 01 08:44:50 crc kubenswrapper[4873]: I1201 08:44:50.304831 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 01 08:44:50 crc kubenswrapper[4873]: I1201 08:44:50.336354 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 01 08:44:50 crc kubenswrapper[4873]: I1201 08:44:50.359824 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 01 08:44:50 crc kubenswrapper[4873]: I1201 08:44:50.359915 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 01 08:44:50 crc kubenswrapper[4873]: I1201 08:44:50.475885 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 01 08:44:50 crc kubenswrapper[4873]: I1201 08:44:50.662522 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 01 08:44:50 crc kubenswrapper[4873]: I1201 08:44:50.736133 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 01 08:44:50 crc kubenswrapper[4873]: I1201 08:44:50.746064 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 01 08:44:50 crc kubenswrapper[4873]: I1201 08:44:50.763561 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 01 08:44:50 crc kubenswrapper[4873]: I1201 08:44:50.844769 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 01 08:44:50 crc kubenswrapper[4873]: I1201 08:44:50.898579 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 01 08:44:50 crc kubenswrapper[4873]: I1201 08:44:50.906767 4873 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 01 08:44:50 crc kubenswrapper[4873]: I1201 08:44:50.915190 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-sg6bg","openshift-kube-apiserver/kube-apiserver-crc"] Dec 01 08:44:50 crc kubenswrapper[4873]: I1201 08:44:50.915284 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-75494747d9-bhtkn","openshift-kube-apiserver/kube-apiserver-crc"] Dec 01 08:44:50 crc kubenswrapper[4873]: E1201 08:44:50.915625 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2404e1f8-203a-4055-8f66-21d8867e7a92" containerName="oauth-openshift" Dec 01 08:44:50 crc kubenswrapper[4873]: I1201 08:44:50.915660 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="2404e1f8-203a-4055-8f66-21d8867e7a92" containerName="oauth-openshift" Dec 01 08:44:50 crc kubenswrapper[4873]: E1201 08:44:50.915702 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ecfbd0a-1599-4102-806c-94067f6235bf" containerName="installer" Dec 01 08:44:50 crc kubenswrapper[4873]: I1201 08:44:50.915720 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ecfbd0a-1599-4102-806c-94067f6235bf" containerName="installer" Dec 01 08:44:50 crc kubenswrapper[4873]: I1201 08:44:50.915925 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ecfbd0a-1599-4102-806c-94067f6235bf" containerName="installer" Dec 01 08:44:50 crc kubenswrapper[4873]: I1201 08:44:50.915963 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="2404e1f8-203a-4055-8f66-21d8867e7a92" containerName="oauth-openshift" Dec 01 08:44:50 crc kubenswrapper[4873]: I1201 08:44:50.916185 4873 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f1c8f149-f188-4759-b2dc-9dc15be1bb13" Dec 01 08:44:50 crc kubenswrapper[4873]: I1201 08:44:50.916256 4873 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f1c8f149-f188-4759-b2dc-9dc15be1bb13" Dec 01 08:44:50 crc kubenswrapper[4873]: I1201 08:44:50.917166 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:44:50 crc kubenswrapper[4873]: I1201 08:44:50.926354 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 01 08:44:50 crc kubenswrapper[4873]: I1201 08:44:50.926426 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 01 08:44:50 crc kubenswrapper[4873]: I1201 08:44:50.927133 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 01 08:44:50 crc kubenswrapper[4873]: I1201 08:44:50.927393 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 01 08:44:50 crc kubenswrapper[4873]: I1201 08:44:50.927465 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 01 08:44:50 crc kubenswrapper[4873]: I1201 08:44:50.928198 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 01 08:44:50 crc kubenswrapper[4873]: I1201 08:44:50.928452 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:44:50 crc kubenswrapper[4873]: I1201 08:44:50.929979 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 01 08:44:50 crc kubenswrapper[4873]: I1201 08:44:50.930326 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 01 08:44:50 crc kubenswrapper[4873]: I1201 08:44:50.930655 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 01 08:44:50 crc kubenswrapper[4873]: I1201 08:44:50.931747 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 01 08:44:50 crc kubenswrapper[4873]: I1201 08:44:50.932377 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 01 08:44:50 crc kubenswrapper[4873]: I1201 08:44:50.934180 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 01 08:44:50 crc kubenswrapper[4873]: I1201 08:44:50.942344 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 01 08:44:50 crc kubenswrapper[4873]: I1201 08:44:50.943694 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 01 08:44:50 crc kubenswrapper[4873]: I1201 08:44:50.979454 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 01 08:44:50 crc kubenswrapper[4873]: I1201 08:44:50.999055 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=15.999033313 podStartE2EDuration="15.999033313s" podCreationTimestamp="2025-12-01 08:44:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:44:50.998250255 +0000 UTC m=+266.900358844" watchObservedRunningTime="2025-12-01 08:44:50.999033313 +0000 UTC m=+266.901141852" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.018310 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c624acad-0f23-48e9-95d1-3cffa69d3b04-v4-0-config-system-cliconfig\") pod \"oauth-openshift-75494747d9-bhtkn\" (UID: \"c624acad-0f23-48e9-95d1-3cffa69d3b04\") " pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.018406 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c624acad-0f23-48e9-95d1-3cffa69d3b04-v4-0-config-system-serving-cert\") pod \"oauth-openshift-75494747d9-bhtkn\" (UID: \"c624acad-0f23-48e9-95d1-3cffa69d3b04\") " pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.018485 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c624acad-0f23-48e9-95d1-3cffa69d3b04-v4-0-config-system-session\") pod \"oauth-openshift-75494747d9-bhtkn\" (UID: \"c624acad-0f23-48e9-95d1-3cffa69d3b04\") " pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.018541 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c624acad-0f23-48e9-95d1-3cffa69d3b04-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-75494747d9-bhtkn\" (UID: \"c624acad-0f23-48e9-95d1-3cffa69d3b04\") " pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.018588 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c624acad-0f23-48e9-95d1-3cffa69d3b04-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-75494747d9-bhtkn\" (UID: \"c624acad-0f23-48e9-95d1-3cffa69d3b04\") " pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.018629 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c624acad-0f23-48e9-95d1-3cffa69d3b04-v4-0-config-system-router-certs\") pod \"oauth-openshift-75494747d9-bhtkn\" (UID: \"c624acad-0f23-48e9-95d1-3cffa69d3b04\") " pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.018673 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c624acad-0f23-48e9-95d1-3cffa69d3b04-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-75494747d9-bhtkn\" (UID: \"c624acad-0f23-48e9-95d1-3cffa69d3b04\") " pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.018728 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c624acad-0f23-48e9-95d1-3cffa69d3b04-audit-dir\") pod \"oauth-openshift-75494747d9-bhtkn\" (UID: \"c624acad-0f23-48e9-95d1-3cffa69d3b04\") " pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.018768 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nmb5\" (UniqueName: \"kubernetes.io/projected/c624acad-0f23-48e9-95d1-3cffa69d3b04-kube-api-access-4nmb5\") pod \"oauth-openshift-75494747d9-bhtkn\" (UID: \"c624acad-0f23-48e9-95d1-3cffa69d3b04\") " pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.018813 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c624acad-0f23-48e9-95d1-3cffa69d3b04-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-75494747d9-bhtkn\" (UID: \"c624acad-0f23-48e9-95d1-3cffa69d3b04\") " pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.018856 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c624acad-0f23-48e9-95d1-3cffa69d3b04-v4-0-config-system-service-ca\") pod \"oauth-openshift-75494747d9-bhtkn\" (UID: \"c624acad-0f23-48e9-95d1-3cffa69d3b04\") " pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.018906 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c624acad-0f23-48e9-95d1-3cffa69d3b04-v4-0-config-user-template-error\") pod \"oauth-openshift-75494747d9-bhtkn\" (UID: \"c624acad-0f23-48e9-95d1-3cffa69d3b04\") " pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.018948 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c624acad-0f23-48e9-95d1-3cffa69d3b04-v4-0-config-user-template-login\") pod \"oauth-openshift-75494747d9-bhtkn\" (UID: \"c624acad-0f23-48e9-95d1-3cffa69d3b04\") " pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.018987 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c624acad-0f23-48e9-95d1-3cffa69d3b04-audit-policies\") pod \"oauth-openshift-75494747d9-bhtkn\" (UID: \"c624acad-0f23-48e9-95d1-3cffa69d3b04\") " pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.080766 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.092205 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.120661 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c624acad-0f23-48e9-95d1-3cffa69d3b04-v4-0-config-system-router-certs\") pod \"oauth-openshift-75494747d9-bhtkn\" (UID: \"c624acad-0f23-48e9-95d1-3cffa69d3b04\") " pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.120739 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c624acad-0f23-48e9-95d1-3cffa69d3b04-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-75494747d9-bhtkn\" (UID: \"c624acad-0f23-48e9-95d1-3cffa69d3b04\") " pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.120783 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c624acad-0f23-48e9-95d1-3cffa69d3b04-audit-dir\") pod \"oauth-openshift-75494747d9-bhtkn\" (UID: \"c624acad-0f23-48e9-95d1-3cffa69d3b04\") " pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.120811 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nmb5\" (UniqueName: \"kubernetes.io/projected/c624acad-0f23-48e9-95d1-3cffa69d3b04-kube-api-access-4nmb5\") pod \"oauth-openshift-75494747d9-bhtkn\" (UID: \"c624acad-0f23-48e9-95d1-3cffa69d3b04\") " pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.120842 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c624acad-0f23-48e9-95d1-3cffa69d3b04-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-75494747d9-bhtkn\" (UID: \"c624acad-0f23-48e9-95d1-3cffa69d3b04\") " pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.120868 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c624acad-0f23-48e9-95d1-3cffa69d3b04-v4-0-config-system-service-ca\") pod \"oauth-openshift-75494747d9-bhtkn\" (UID: \"c624acad-0f23-48e9-95d1-3cffa69d3b04\") " pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.120900 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c624acad-0f23-48e9-95d1-3cffa69d3b04-v4-0-config-user-template-error\") pod \"oauth-openshift-75494747d9-bhtkn\" (UID: \"c624acad-0f23-48e9-95d1-3cffa69d3b04\") " pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.120926 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c624acad-0f23-48e9-95d1-3cffa69d3b04-audit-policies\") pod \"oauth-openshift-75494747d9-bhtkn\" (UID: \"c624acad-0f23-48e9-95d1-3cffa69d3b04\") " pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.120949 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c624acad-0f23-48e9-95d1-3cffa69d3b04-v4-0-config-user-template-login\") pod \"oauth-openshift-75494747d9-bhtkn\" (UID: \"c624acad-0f23-48e9-95d1-3cffa69d3b04\") " pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.121028 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c624acad-0f23-48e9-95d1-3cffa69d3b04-v4-0-config-system-cliconfig\") pod \"oauth-openshift-75494747d9-bhtkn\" (UID: \"c624acad-0f23-48e9-95d1-3cffa69d3b04\") " pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.121057 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c624acad-0f23-48e9-95d1-3cffa69d3b04-v4-0-config-system-serving-cert\") pod \"oauth-openshift-75494747d9-bhtkn\" (UID: \"c624acad-0f23-48e9-95d1-3cffa69d3b04\") " pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.121106 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c624acad-0f23-48e9-95d1-3cffa69d3b04-v4-0-config-system-session\") pod \"oauth-openshift-75494747d9-bhtkn\" (UID: \"c624acad-0f23-48e9-95d1-3cffa69d3b04\") " pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.121136 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c624acad-0f23-48e9-95d1-3cffa69d3b04-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-75494747d9-bhtkn\" (UID: \"c624acad-0f23-48e9-95d1-3cffa69d3b04\") " pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.121171 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c624acad-0f23-48e9-95d1-3cffa69d3b04-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-75494747d9-bhtkn\" (UID: \"c624acad-0f23-48e9-95d1-3cffa69d3b04\") " pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.122392 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c624acad-0f23-48e9-95d1-3cffa69d3b04-v4-0-config-system-service-ca\") pod \"oauth-openshift-75494747d9-bhtkn\" (UID: \"c624acad-0f23-48e9-95d1-3cffa69d3b04\") " pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.122484 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c624acad-0f23-48e9-95d1-3cffa69d3b04-audit-policies\") pod \"oauth-openshift-75494747d9-bhtkn\" (UID: \"c624acad-0f23-48e9-95d1-3cffa69d3b04\") " pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.123145 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c624acad-0f23-48e9-95d1-3cffa69d3b04-v4-0-config-system-cliconfig\") pod \"oauth-openshift-75494747d9-bhtkn\" (UID: \"c624acad-0f23-48e9-95d1-3cffa69d3b04\") " pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.120964 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c624acad-0f23-48e9-95d1-3cffa69d3b04-audit-dir\") pod \"oauth-openshift-75494747d9-bhtkn\" (UID: \"c624acad-0f23-48e9-95d1-3cffa69d3b04\") " pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.124626 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c624acad-0f23-48e9-95d1-3cffa69d3b04-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-75494747d9-bhtkn\" (UID: \"c624acad-0f23-48e9-95d1-3cffa69d3b04\") " pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.129097 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c624acad-0f23-48e9-95d1-3cffa69d3b04-v4-0-config-system-router-certs\") pod \"oauth-openshift-75494747d9-bhtkn\" (UID: \"c624acad-0f23-48e9-95d1-3cffa69d3b04\") " pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.129576 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c624acad-0f23-48e9-95d1-3cffa69d3b04-v4-0-config-user-template-error\") pod \"oauth-openshift-75494747d9-bhtkn\" (UID: \"c624acad-0f23-48e9-95d1-3cffa69d3b04\") " pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.129784 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c624acad-0f23-48e9-95d1-3cffa69d3b04-v4-0-config-user-template-login\") pod \"oauth-openshift-75494747d9-bhtkn\" (UID: \"c624acad-0f23-48e9-95d1-3cffa69d3b04\") " pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.130128 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c624acad-0f23-48e9-95d1-3cffa69d3b04-v4-0-config-system-serving-cert\") pod \"oauth-openshift-75494747d9-bhtkn\" (UID: \"c624acad-0f23-48e9-95d1-3cffa69d3b04\") " pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.130430 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c624acad-0f23-48e9-95d1-3cffa69d3b04-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-75494747d9-bhtkn\" (UID: \"c624acad-0f23-48e9-95d1-3cffa69d3b04\") " pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.131429 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c624acad-0f23-48e9-95d1-3cffa69d3b04-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-75494747d9-bhtkn\" (UID: \"c624acad-0f23-48e9-95d1-3cffa69d3b04\") " pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.131637 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c624acad-0f23-48e9-95d1-3cffa69d3b04-v4-0-config-system-session\") pod \"oauth-openshift-75494747d9-bhtkn\" (UID: \"c624acad-0f23-48e9-95d1-3cffa69d3b04\") " pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.141067 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nmb5\" (UniqueName: \"kubernetes.io/projected/c624acad-0f23-48e9-95d1-3cffa69d3b04-kube-api-access-4nmb5\") pod \"oauth-openshift-75494747d9-bhtkn\" (UID: \"c624acad-0f23-48e9-95d1-3cffa69d3b04\") " pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.144843 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c624acad-0f23-48e9-95d1-3cffa69d3b04-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-75494747d9-bhtkn\" (UID: \"c624acad-0f23-48e9-95d1-3cffa69d3b04\") " pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.154476 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.250482 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.305185 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.309276 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.337055 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.563081 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.631874 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.652250 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.661139 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.662143 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.728003 4873 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.828509 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 01 08:44:51 crc kubenswrapper[4873]: I1201 08:44:51.955075 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 01 08:44:52 crc kubenswrapper[4873]: I1201 08:44:52.007157 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 01 08:44:52 crc kubenswrapper[4873]: I1201 08:44:52.045315 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 01 08:44:52 crc kubenswrapper[4873]: I1201 08:44:52.090072 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 01 08:44:52 crc kubenswrapper[4873]: I1201 08:44:52.166086 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 01 08:44:52 crc kubenswrapper[4873]: I1201 08:44:52.173494 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 01 08:44:52 crc kubenswrapper[4873]: I1201 08:44:52.223428 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 01 08:44:52 crc kubenswrapper[4873]: I1201 08:44:52.231616 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 01 08:44:52 crc kubenswrapper[4873]: I1201 08:44:52.244881 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 01 08:44:52 crc kubenswrapper[4873]: I1201 08:44:52.277574 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 01 08:44:52 crc kubenswrapper[4873]: I1201 08:44:52.426930 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 01 08:44:52 crc kubenswrapper[4873]: I1201 08:44:52.440839 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2404e1f8-203a-4055-8f66-21d8867e7a92" path="/var/lib/kubelet/pods/2404e1f8-203a-4055-8f66-21d8867e7a92/volumes" Dec 01 08:44:52 crc kubenswrapper[4873]: I1201 08:44:52.445833 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 01 08:44:52 crc kubenswrapper[4873]: I1201 08:44:52.478880 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 01 08:44:52 crc kubenswrapper[4873]: I1201 08:44:52.549122 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 01 08:44:52 crc kubenswrapper[4873]: I1201 08:44:52.729452 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 01 08:44:52 crc kubenswrapper[4873]: I1201 08:44:52.830089 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 01 08:44:52 crc kubenswrapper[4873]: I1201 08:44:52.911754 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 01 08:44:52 crc kubenswrapper[4873]: I1201 08:44:52.928003 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 01 08:44:52 crc kubenswrapper[4873]: I1201 08:44:52.998041 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 01 08:44:53 crc kubenswrapper[4873]: I1201 08:44:53.045600 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 01 08:44:53 crc kubenswrapper[4873]: I1201 08:44:53.108767 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 01 08:44:53 crc kubenswrapper[4873]: I1201 08:44:53.126891 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 01 08:44:53 crc kubenswrapper[4873]: I1201 08:44:53.205549 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 01 08:44:53 crc kubenswrapper[4873]: I1201 08:44:53.259243 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 01 08:44:53 crc kubenswrapper[4873]: I1201 08:44:53.279062 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 01 08:44:53 crc kubenswrapper[4873]: I1201 08:44:53.279479 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 01 08:44:53 crc kubenswrapper[4873]: I1201 08:44:53.344246 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 01 08:44:53 crc kubenswrapper[4873]: I1201 08:44:53.356235 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 01 08:44:53 crc kubenswrapper[4873]: I1201 08:44:53.406333 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 01 08:44:53 crc kubenswrapper[4873]: I1201 08:44:53.406827 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 01 08:44:53 crc kubenswrapper[4873]: I1201 08:44:53.427355 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 01 08:44:53 crc kubenswrapper[4873]: I1201 08:44:53.428092 4873 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 01 08:44:53 crc kubenswrapper[4873]: I1201 08:44:53.592979 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 01 08:44:53 crc kubenswrapper[4873]: I1201 08:44:53.632603 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 01 08:44:53 crc kubenswrapper[4873]: I1201 08:44:53.649501 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 01 08:44:53 crc kubenswrapper[4873]: I1201 08:44:53.658192 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 01 08:44:53 crc kubenswrapper[4873]: I1201 08:44:53.773597 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 01 08:44:53 crc kubenswrapper[4873]: I1201 08:44:53.806763 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 01 08:44:53 crc kubenswrapper[4873]: I1201 08:44:53.850351 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 01 08:44:53 crc kubenswrapper[4873]: I1201 08:44:53.929289 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 01 08:44:53 crc kubenswrapper[4873]: I1201 08:44:53.962582 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 01 08:44:54 crc kubenswrapper[4873]: I1201 08:44:54.001986 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 01 08:44:54 crc kubenswrapper[4873]: I1201 08:44:54.050488 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 01 08:44:54 crc kubenswrapper[4873]: I1201 08:44:54.078406 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 01 08:44:54 crc kubenswrapper[4873]: I1201 08:44:54.112720 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 01 08:44:54 crc kubenswrapper[4873]: E1201 08:44:54.136712 4873 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 01 08:44:54 crc kubenswrapper[4873]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-75494747d9-bhtkn_openshift-authentication_c624acad-0f23-48e9-95d1-3cffa69d3b04_0(9b9d736a9a7eb85e579eee6ab8adc4c0bcd4b0371e5e73e1161a7ddd411440be): error adding pod openshift-authentication_oauth-openshift-75494747d9-bhtkn to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"9b9d736a9a7eb85e579eee6ab8adc4c0bcd4b0371e5e73e1161a7ddd411440be" Netns:"/var/run/netns/ab0c659e-7ce7-4e6d-b63c-c1295aaba436" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-75494747d9-bhtkn;K8S_POD_INFRA_CONTAINER_ID=9b9d736a9a7eb85e579eee6ab8adc4c0bcd4b0371e5e73e1161a7ddd411440be;K8S_POD_UID=c624acad-0f23-48e9-95d1-3cffa69d3b04" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-75494747d9-bhtkn] networking: Multus: [openshift-authentication/oauth-openshift-75494747d9-bhtkn/c624acad-0f23-48e9-95d1-3cffa69d3b04]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-75494747d9-bhtkn in out of cluster comm: pod "oauth-openshift-75494747d9-bhtkn" not found Dec 01 08:44:54 crc kubenswrapper[4873]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 01 08:44:54 crc kubenswrapper[4873]: > Dec 01 08:44:54 crc kubenswrapper[4873]: E1201 08:44:54.136831 4873 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 01 08:44:54 crc kubenswrapper[4873]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-75494747d9-bhtkn_openshift-authentication_c624acad-0f23-48e9-95d1-3cffa69d3b04_0(9b9d736a9a7eb85e579eee6ab8adc4c0bcd4b0371e5e73e1161a7ddd411440be): error adding pod openshift-authentication_oauth-openshift-75494747d9-bhtkn to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"9b9d736a9a7eb85e579eee6ab8adc4c0bcd4b0371e5e73e1161a7ddd411440be" Netns:"/var/run/netns/ab0c659e-7ce7-4e6d-b63c-c1295aaba436" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-75494747d9-bhtkn;K8S_POD_INFRA_CONTAINER_ID=9b9d736a9a7eb85e579eee6ab8adc4c0bcd4b0371e5e73e1161a7ddd411440be;K8S_POD_UID=c624acad-0f23-48e9-95d1-3cffa69d3b04" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-75494747d9-bhtkn] networking: Multus: [openshift-authentication/oauth-openshift-75494747d9-bhtkn/c624acad-0f23-48e9-95d1-3cffa69d3b04]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-75494747d9-bhtkn in out of cluster comm: pod "oauth-openshift-75494747d9-bhtkn" not found Dec 01 08:44:54 crc kubenswrapper[4873]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 01 08:44:54 crc kubenswrapper[4873]: > pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:44:54 crc kubenswrapper[4873]: E1201 08:44:54.136856 4873 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Dec 01 08:44:54 crc kubenswrapper[4873]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-75494747d9-bhtkn_openshift-authentication_c624acad-0f23-48e9-95d1-3cffa69d3b04_0(9b9d736a9a7eb85e579eee6ab8adc4c0bcd4b0371e5e73e1161a7ddd411440be): error adding pod openshift-authentication_oauth-openshift-75494747d9-bhtkn to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"9b9d736a9a7eb85e579eee6ab8adc4c0bcd4b0371e5e73e1161a7ddd411440be" Netns:"/var/run/netns/ab0c659e-7ce7-4e6d-b63c-c1295aaba436" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-75494747d9-bhtkn;K8S_POD_INFRA_CONTAINER_ID=9b9d736a9a7eb85e579eee6ab8adc4c0bcd4b0371e5e73e1161a7ddd411440be;K8S_POD_UID=c624acad-0f23-48e9-95d1-3cffa69d3b04" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-75494747d9-bhtkn] networking: Multus: [openshift-authentication/oauth-openshift-75494747d9-bhtkn/c624acad-0f23-48e9-95d1-3cffa69d3b04]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-75494747d9-bhtkn in out of cluster comm: pod "oauth-openshift-75494747d9-bhtkn" not found Dec 01 08:44:54 crc kubenswrapper[4873]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 01 08:44:54 crc kubenswrapper[4873]: > pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:44:54 crc kubenswrapper[4873]: E1201 08:44:54.136941 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"oauth-openshift-75494747d9-bhtkn_openshift-authentication(c624acad-0f23-48e9-95d1-3cffa69d3b04)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"oauth-openshift-75494747d9-bhtkn_openshift-authentication(c624acad-0f23-48e9-95d1-3cffa69d3b04)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-75494747d9-bhtkn_openshift-authentication_c624acad-0f23-48e9-95d1-3cffa69d3b04_0(9b9d736a9a7eb85e579eee6ab8adc4c0bcd4b0371e5e73e1161a7ddd411440be): error adding pod openshift-authentication_oauth-openshift-75494747d9-bhtkn to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"9b9d736a9a7eb85e579eee6ab8adc4c0bcd4b0371e5e73e1161a7ddd411440be\\\" Netns:\\\"/var/run/netns/ab0c659e-7ce7-4e6d-b63c-c1295aaba436\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-75494747d9-bhtkn;K8S_POD_INFRA_CONTAINER_ID=9b9d736a9a7eb85e579eee6ab8adc4c0bcd4b0371e5e73e1161a7ddd411440be;K8S_POD_UID=c624acad-0f23-48e9-95d1-3cffa69d3b04\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-75494747d9-bhtkn] networking: Multus: [openshift-authentication/oauth-openshift-75494747d9-bhtkn/c624acad-0f23-48e9-95d1-3cffa69d3b04]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-75494747d9-bhtkn in out of cluster comm: pod \\\"oauth-openshift-75494747d9-bhtkn\\\" not found\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" podUID="c624acad-0f23-48e9-95d1-3cffa69d3b04" Dec 01 08:44:54 crc kubenswrapper[4873]: I1201 08:44:54.149198 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 01 08:44:54 crc kubenswrapper[4873]: I1201 08:44:54.168901 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 01 08:44:54 crc kubenswrapper[4873]: I1201 08:44:54.181430 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 01 08:44:54 crc kubenswrapper[4873]: I1201 08:44:54.254926 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 01 08:44:54 crc kubenswrapper[4873]: I1201 08:44:54.260201 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 01 08:44:54 crc kubenswrapper[4873]: I1201 08:44:54.402684 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 01 08:44:54 crc kubenswrapper[4873]: I1201 08:44:54.417572 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 01 08:44:54 crc kubenswrapper[4873]: I1201 08:44:54.474431 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 01 08:44:54 crc kubenswrapper[4873]: I1201 08:44:54.634240 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 01 08:44:54 crc kubenswrapper[4873]: I1201 08:44:54.641225 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 01 08:44:54 crc kubenswrapper[4873]: I1201 08:44:54.662327 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 01 08:44:54 crc kubenswrapper[4873]: I1201 08:44:54.665626 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 01 08:44:54 crc kubenswrapper[4873]: I1201 08:44:54.724363 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 01 08:44:54 crc kubenswrapper[4873]: I1201 08:44:54.777740 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 01 08:44:54 crc kubenswrapper[4873]: I1201 08:44:54.801211 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 01 08:44:54 crc kubenswrapper[4873]: I1201 08:44:54.865049 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 01 08:44:54 crc kubenswrapper[4873]: I1201 08:44:54.921635 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 01 08:44:54 crc kubenswrapper[4873]: I1201 08:44:54.997143 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 01 08:44:55 crc kubenswrapper[4873]: I1201 08:44:55.133205 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 01 08:44:55 crc kubenswrapper[4873]: I1201 08:44:55.157876 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 01 08:44:55 crc kubenswrapper[4873]: I1201 08:44:55.314104 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 01 08:44:55 crc kubenswrapper[4873]: I1201 08:44:55.319873 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 01 08:44:55 crc kubenswrapper[4873]: I1201 08:44:55.330611 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 01 08:44:55 crc kubenswrapper[4873]: I1201 08:44:55.416385 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 01 08:44:55 crc kubenswrapper[4873]: I1201 08:44:55.438497 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 01 08:44:55 crc kubenswrapper[4873]: I1201 08:44:55.443280 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 01 08:44:55 crc kubenswrapper[4873]: I1201 08:44:55.558169 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 01 08:44:55 crc kubenswrapper[4873]: I1201 08:44:55.665287 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 01 08:44:55 crc kubenswrapper[4873]: I1201 08:44:55.665419 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 01 08:44:55 crc kubenswrapper[4873]: I1201 08:44:55.791887 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 01 08:44:55 crc kubenswrapper[4873]: I1201 08:44:55.825790 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 01 08:44:55 crc kubenswrapper[4873]: I1201 08:44:55.964495 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 01 08:44:55 crc kubenswrapper[4873]: I1201 08:44:55.989954 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 01 08:44:56 crc kubenswrapper[4873]: I1201 08:44:56.036141 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 01 08:44:56 crc kubenswrapper[4873]: I1201 08:44:56.047118 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 01 08:44:56 crc kubenswrapper[4873]: I1201 08:44:56.054532 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 01 08:44:56 crc kubenswrapper[4873]: I1201 08:44:56.180243 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 01 08:44:56 crc kubenswrapper[4873]: I1201 08:44:56.212931 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 01 08:44:56 crc kubenswrapper[4873]: I1201 08:44:56.315232 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 01 08:44:56 crc kubenswrapper[4873]: I1201 08:44:56.340665 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 01 08:44:56 crc kubenswrapper[4873]: I1201 08:44:56.359263 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 01 08:44:56 crc kubenswrapper[4873]: I1201 08:44:56.385400 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 01 08:44:56 crc kubenswrapper[4873]: I1201 08:44:56.424186 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 01 08:44:56 crc kubenswrapper[4873]: I1201 08:44:56.448234 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 01 08:44:56 crc kubenswrapper[4873]: I1201 08:44:56.458299 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 01 08:44:56 crc kubenswrapper[4873]: I1201 08:44:56.509429 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 01 08:44:56 crc kubenswrapper[4873]: I1201 08:44:56.549611 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 01 08:44:56 crc kubenswrapper[4873]: I1201 08:44:56.591484 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 01 08:44:56 crc kubenswrapper[4873]: I1201 08:44:56.731813 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 01 08:44:56 crc kubenswrapper[4873]: I1201 08:44:56.734307 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 01 08:44:56 crc kubenswrapper[4873]: I1201 08:44:56.735381 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 01 08:44:56 crc kubenswrapper[4873]: I1201 08:44:56.763717 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 01 08:44:56 crc kubenswrapper[4873]: I1201 08:44:56.783126 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 01 08:44:56 crc kubenswrapper[4873]: I1201 08:44:56.790932 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 01 08:44:56 crc kubenswrapper[4873]: I1201 08:44:56.842752 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 01 08:44:56 crc kubenswrapper[4873]: I1201 08:44:56.885607 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 01 08:44:57 crc kubenswrapper[4873]: I1201 08:44:57.033183 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 01 08:44:57 crc kubenswrapper[4873]: I1201 08:44:57.044451 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 01 08:44:57 crc kubenswrapper[4873]: I1201 08:44:57.049907 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 01 08:44:57 crc kubenswrapper[4873]: I1201 08:44:57.070685 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 01 08:44:57 crc kubenswrapper[4873]: I1201 08:44:57.078405 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 01 08:44:57 crc kubenswrapper[4873]: I1201 08:44:57.109762 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 01 08:44:57 crc kubenswrapper[4873]: I1201 08:44:57.144357 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 01 08:44:57 crc kubenswrapper[4873]: I1201 08:44:57.214759 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 01 08:44:57 crc kubenswrapper[4873]: I1201 08:44:57.240220 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 01 08:44:57 crc kubenswrapper[4873]: I1201 08:44:57.240750 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 01 08:44:57 crc kubenswrapper[4873]: I1201 08:44:57.311196 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 01 08:44:57 crc kubenswrapper[4873]: I1201 08:44:57.334213 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 01 08:44:57 crc kubenswrapper[4873]: I1201 08:44:57.393809 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 01 08:44:57 crc kubenswrapper[4873]: I1201 08:44:57.418249 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 01 08:44:57 crc kubenswrapper[4873]: I1201 08:44:57.422671 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 01 08:44:57 crc kubenswrapper[4873]: I1201 08:44:57.539229 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 01 08:44:57 crc kubenswrapper[4873]: I1201 08:44:57.575848 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 01 08:44:57 crc kubenswrapper[4873]: I1201 08:44:57.615253 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 01 08:44:57 crc kubenswrapper[4873]: I1201 08:44:57.615394 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 01 08:44:57 crc kubenswrapper[4873]: I1201 08:44:57.677050 4873 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 01 08:44:57 crc kubenswrapper[4873]: I1201 08:44:57.677355 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://457fa84c82e055fb574e97e03a4a78e5ed995026a8c5549a445806c28cb03627" gracePeriod=5 Dec 01 08:44:57 crc kubenswrapper[4873]: I1201 08:44:57.710961 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 01 08:44:57 crc kubenswrapper[4873]: I1201 08:44:57.713608 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 01 08:44:57 crc kubenswrapper[4873]: I1201 08:44:57.821074 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 01 08:44:57 crc kubenswrapper[4873]: I1201 08:44:57.945939 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 01 08:44:57 crc kubenswrapper[4873]: I1201 08:44:57.954116 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 01 08:44:58 crc kubenswrapper[4873]: I1201 08:44:58.008416 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 01 08:44:58 crc kubenswrapper[4873]: I1201 08:44:58.013232 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 01 08:44:58 crc kubenswrapper[4873]: I1201 08:44:58.203338 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 01 08:44:58 crc kubenswrapper[4873]: I1201 08:44:58.215620 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 01 08:44:58 crc kubenswrapper[4873]: I1201 08:44:58.245091 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 01 08:44:58 crc kubenswrapper[4873]: I1201 08:44:58.264610 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 01 08:44:58 crc kubenswrapper[4873]: I1201 08:44:58.485080 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 01 08:44:58 crc kubenswrapper[4873]: I1201 08:44:58.517772 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 01 08:44:58 crc kubenswrapper[4873]: I1201 08:44:58.691321 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 01 08:44:58 crc kubenswrapper[4873]: I1201 08:44:58.792678 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 01 08:44:58 crc kubenswrapper[4873]: I1201 08:44:58.825925 4873 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 01 08:44:58 crc kubenswrapper[4873]: I1201 08:44:58.941473 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 01 08:44:58 crc kubenswrapper[4873]: I1201 08:44:58.997616 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 01 08:44:59 crc kubenswrapper[4873]: I1201 08:44:59.125712 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 01 08:44:59 crc kubenswrapper[4873]: I1201 08:44:59.266468 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 01 08:44:59 crc kubenswrapper[4873]: I1201 08:44:59.408139 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 01 08:44:59 crc kubenswrapper[4873]: I1201 08:44:59.573362 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 01 08:44:59 crc kubenswrapper[4873]: I1201 08:44:59.704494 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 01 08:44:59 crc kubenswrapper[4873]: I1201 08:44:59.728768 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 01 08:44:59 crc kubenswrapper[4873]: I1201 08:44:59.893533 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 01 08:45:00 crc kubenswrapper[4873]: I1201 08:45:00.007433 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 01 08:45:00 crc kubenswrapper[4873]: I1201 08:45:00.171925 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409645-vfsz2"] Dec 01 08:45:00 crc kubenswrapper[4873]: E1201 08:45:00.172269 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 01 08:45:00 crc kubenswrapper[4873]: I1201 08:45:00.172285 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 01 08:45:00 crc kubenswrapper[4873]: I1201 08:45:00.172399 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 01 08:45:00 crc kubenswrapper[4873]: I1201 08:45:00.172943 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-vfsz2" Dec 01 08:45:00 crc kubenswrapper[4873]: I1201 08:45:00.178395 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 08:45:00 crc kubenswrapper[4873]: I1201 08:45:00.179279 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 08:45:00 crc kubenswrapper[4873]: I1201 08:45:00.195565 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409645-vfsz2"] Dec 01 08:45:00 crc kubenswrapper[4873]: I1201 08:45:00.253079 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f6d9174b-c4b0-4176-9e95-291c04608323-secret-volume\") pod \"collect-profiles-29409645-vfsz2\" (UID: \"f6d9174b-c4b0-4176-9e95-291c04608323\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-vfsz2" Dec 01 08:45:00 crc kubenswrapper[4873]: I1201 08:45:00.253254 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bm8s\" (UniqueName: \"kubernetes.io/projected/f6d9174b-c4b0-4176-9e95-291c04608323-kube-api-access-4bm8s\") pod \"collect-profiles-29409645-vfsz2\" (UID: \"f6d9174b-c4b0-4176-9e95-291c04608323\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-vfsz2" Dec 01 08:45:00 crc kubenswrapper[4873]: I1201 08:45:00.253426 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f6d9174b-c4b0-4176-9e95-291c04608323-config-volume\") pod \"collect-profiles-29409645-vfsz2\" (UID: \"f6d9174b-c4b0-4176-9e95-291c04608323\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-vfsz2" Dec 01 08:45:00 crc kubenswrapper[4873]: I1201 08:45:00.344248 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 01 08:45:00 crc kubenswrapper[4873]: I1201 08:45:00.354587 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f6d9174b-c4b0-4176-9e95-291c04608323-config-volume\") pod \"collect-profiles-29409645-vfsz2\" (UID: \"f6d9174b-c4b0-4176-9e95-291c04608323\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-vfsz2" Dec 01 08:45:00 crc kubenswrapper[4873]: I1201 08:45:00.354642 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f6d9174b-c4b0-4176-9e95-291c04608323-secret-volume\") pod \"collect-profiles-29409645-vfsz2\" (UID: \"f6d9174b-c4b0-4176-9e95-291c04608323\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-vfsz2" Dec 01 08:45:00 crc kubenswrapper[4873]: I1201 08:45:00.354712 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bm8s\" (UniqueName: \"kubernetes.io/projected/f6d9174b-c4b0-4176-9e95-291c04608323-kube-api-access-4bm8s\") pod \"collect-profiles-29409645-vfsz2\" (UID: \"f6d9174b-c4b0-4176-9e95-291c04608323\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-vfsz2" Dec 01 08:45:00 crc kubenswrapper[4873]: I1201 08:45:00.355771 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f6d9174b-c4b0-4176-9e95-291c04608323-config-volume\") pod \"collect-profiles-29409645-vfsz2\" (UID: \"f6d9174b-c4b0-4176-9e95-291c04608323\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-vfsz2" Dec 01 08:45:00 crc kubenswrapper[4873]: I1201 08:45:00.373701 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bm8s\" (UniqueName: \"kubernetes.io/projected/f6d9174b-c4b0-4176-9e95-291c04608323-kube-api-access-4bm8s\") pod \"collect-profiles-29409645-vfsz2\" (UID: \"f6d9174b-c4b0-4176-9e95-291c04608323\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-vfsz2" Dec 01 08:45:00 crc kubenswrapper[4873]: I1201 08:45:00.375164 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f6d9174b-c4b0-4176-9e95-291c04608323-secret-volume\") pod \"collect-profiles-29409645-vfsz2\" (UID: \"f6d9174b-c4b0-4176-9e95-291c04608323\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-vfsz2" Dec 01 08:45:00 crc kubenswrapper[4873]: I1201 08:45:00.503461 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-vfsz2" Dec 01 08:45:00 crc kubenswrapper[4873]: I1201 08:45:00.590411 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 01 08:45:00 crc kubenswrapper[4873]: I1201 08:45:00.596388 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 01 08:45:00 crc kubenswrapper[4873]: I1201 08:45:00.674003 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 01 08:45:00 crc kubenswrapper[4873]: I1201 08:45:00.691706 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409645-vfsz2"] Dec 01 08:45:00 crc kubenswrapper[4873]: I1201 08:45:00.775287 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 01 08:45:00 crc kubenswrapper[4873]: I1201 08:45:00.843168 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 01 08:45:00 crc kubenswrapper[4873]: I1201 08:45:00.844638 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 01 08:45:00 crc kubenswrapper[4873]: I1201 08:45:00.985843 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 01 08:45:01 crc kubenswrapper[4873]: I1201 08:45:01.209551 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 01 08:45:01 crc kubenswrapper[4873]: I1201 08:45:01.266678 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 01 08:45:01 crc kubenswrapper[4873]: I1201 08:45:01.390305 4873 generic.go:334] "Generic (PLEG): container finished" podID="f6d9174b-c4b0-4176-9e95-291c04608323" containerID="dd3d81fd19b458c0b28f082da23039ebaaab818011a7afc53a450012b4b8c01b" exitCode=0 Dec 01 08:45:01 crc kubenswrapper[4873]: I1201 08:45:01.390405 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-vfsz2" event={"ID":"f6d9174b-c4b0-4176-9e95-291c04608323","Type":"ContainerDied","Data":"dd3d81fd19b458c0b28f082da23039ebaaab818011a7afc53a450012b4b8c01b"} Dec 01 08:45:01 crc kubenswrapper[4873]: I1201 08:45:01.390475 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-vfsz2" event={"ID":"f6d9174b-c4b0-4176-9e95-291c04608323","Type":"ContainerStarted","Data":"b880b5cd87d3e6daabcfee3aafd969b7126a7fb5c3c8d1d1444aa9093b3c612a"} Dec 01 08:45:01 crc kubenswrapper[4873]: I1201 08:45:01.672382 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 01 08:45:01 crc kubenswrapper[4873]: I1201 08:45:01.780957 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 01 08:45:02 crc kubenswrapper[4873]: I1201 08:45:02.547382 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 01 08:45:02 crc kubenswrapper[4873]: I1201 08:45:02.612325 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-vfsz2" Dec 01 08:45:02 crc kubenswrapper[4873]: I1201 08:45:02.689626 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f6d9174b-c4b0-4176-9e95-291c04608323-secret-volume\") pod \"f6d9174b-c4b0-4176-9e95-291c04608323\" (UID: \"f6d9174b-c4b0-4176-9e95-291c04608323\") " Dec 01 08:45:02 crc kubenswrapper[4873]: I1201 08:45:02.689706 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4bm8s\" (UniqueName: \"kubernetes.io/projected/f6d9174b-c4b0-4176-9e95-291c04608323-kube-api-access-4bm8s\") pod \"f6d9174b-c4b0-4176-9e95-291c04608323\" (UID: \"f6d9174b-c4b0-4176-9e95-291c04608323\") " Dec 01 08:45:02 crc kubenswrapper[4873]: I1201 08:45:02.689768 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f6d9174b-c4b0-4176-9e95-291c04608323-config-volume\") pod \"f6d9174b-c4b0-4176-9e95-291c04608323\" (UID: \"f6d9174b-c4b0-4176-9e95-291c04608323\") " Dec 01 08:45:02 crc kubenswrapper[4873]: I1201 08:45:02.690914 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6d9174b-c4b0-4176-9e95-291c04608323-config-volume" (OuterVolumeSpecName: "config-volume") pod "f6d9174b-c4b0-4176-9e95-291c04608323" (UID: "f6d9174b-c4b0-4176-9e95-291c04608323"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:45:02 crc kubenswrapper[4873]: I1201 08:45:02.694938 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6d9174b-c4b0-4176-9e95-291c04608323-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f6d9174b-c4b0-4176-9e95-291c04608323" (UID: "f6d9174b-c4b0-4176-9e95-291c04608323"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:45:02 crc kubenswrapper[4873]: I1201 08:45:02.696262 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6d9174b-c4b0-4176-9e95-291c04608323-kube-api-access-4bm8s" (OuterVolumeSpecName: "kube-api-access-4bm8s") pod "f6d9174b-c4b0-4176-9e95-291c04608323" (UID: "f6d9174b-c4b0-4176-9e95-291c04608323"). InnerVolumeSpecName "kube-api-access-4bm8s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:45:02 crc kubenswrapper[4873]: I1201 08:45:02.791474 4873 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f6d9174b-c4b0-4176-9e95-291c04608323-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:02 crc kubenswrapper[4873]: I1201 08:45:02.791523 4873 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f6d9174b-c4b0-4176-9e95-291c04608323-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:02 crc kubenswrapper[4873]: I1201 08:45:02.791536 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4bm8s\" (UniqueName: \"kubernetes.io/projected/f6d9174b-c4b0-4176-9e95-291c04608323-kube-api-access-4bm8s\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:02 crc kubenswrapper[4873]: I1201 08:45:02.828973 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 01 08:45:02 crc kubenswrapper[4873]: I1201 08:45:02.829083 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:45:02 crc kubenswrapper[4873]: I1201 08:45:02.892148 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 01 08:45:02 crc kubenswrapper[4873]: I1201 08:45:02.892318 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:45:02 crc kubenswrapper[4873]: I1201 08:45:02.892350 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 01 08:45:02 crc kubenswrapper[4873]: I1201 08:45:02.892380 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 01 08:45:02 crc kubenswrapper[4873]: I1201 08:45:02.892401 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 01 08:45:02 crc kubenswrapper[4873]: I1201 08:45:02.892432 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 01 08:45:02 crc kubenswrapper[4873]: I1201 08:45:02.892481 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:45:02 crc kubenswrapper[4873]: I1201 08:45:02.892509 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:45:02 crc kubenswrapper[4873]: I1201 08:45:02.892607 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:45:02 crc kubenswrapper[4873]: I1201 08:45:02.892820 4873 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:02 crc kubenswrapper[4873]: I1201 08:45:02.892837 4873 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:02 crc kubenswrapper[4873]: I1201 08:45:02.892846 4873 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:02 crc kubenswrapper[4873]: I1201 08:45:02.892856 4873 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:02 crc kubenswrapper[4873]: I1201 08:45:02.897959 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:45:02 crc kubenswrapper[4873]: I1201 08:45:02.995000 4873 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:03 crc kubenswrapper[4873]: I1201 08:45:03.385763 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 01 08:45:03 crc kubenswrapper[4873]: I1201 08:45:03.404211 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-vfsz2" Dec 01 08:45:03 crc kubenswrapper[4873]: I1201 08:45:03.404153 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-vfsz2" event={"ID":"f6d9174b-c4b0-4176-9e95-291c04608323","Type":"ContainerDied","Data":"b880b5cd87d3e6daabcfee3aafd969b7126a7fb5c3c8d1d1444aa9093b3c612a"} Dec 01 08:45:03 crc kubenswrapper[4873]: I1201 08:45:03.404287 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b880b5cd87d3e6daabcfee3aafd969b7126a7fb5c3c8d1d1444aa9093b3c612a" Dec 01 08:45:03 crc kubenswrapper[4873]: I1201 08:45:03.407307 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 01 08:45:03 crc kubenswrapper[4873]: I1201 08:45:03.407367 4873 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="457fa84c82e055fb574e97e03a4a78e5ed995026a8c5549a445806c28cb03627" exitCode=137 Dec 01 08:45:03 crc kubenswrapper[4873]: I1201 08:45:03.407418 4873 scope.go:117] "RemoveContainer" containerID="457fa84c82e055fb574e97e03a4a78e5ed995026a8c5549a445806c28cb03627" Dec 01 08:45:03 crc kubenswrapper[4873]: I1201 08:45:03.407607 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:45:03 crc kubenswrapper[4873]: I1201 08:45:03.429347 4873 scope.go:117] "RemoveContainer" containerID="457fa84c82e055fb574e97e03a4a78e5ed995026a8c5549a445806c28cb03627" Dec 01 08:45:03 crc kubenswrapper[4873]: E1201 08:45:03.429990 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"457fa84c82e055fb574e97e03a4a78e5ed995026a8c5549a445806c28cb03627\": container with ID starting with 457fa84c82e055fb574e97e03a4a78e5ed995026a8c5549a445806c28cb03627 not found: ID does not exist" containerID="457fa84c82e055fb574e97e03a4a78e5ed995026a8c5549a445806c28cb03627" Dec 01 08:45:03 crc kubenswrapper[4873]: I1201 08:45:03.430044 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"457fa84c82e055fb574e97e03a4a78e5ed995026a8c5549a445806c28cb03627"} err="failed to get container status \"457fa84c82e055fb574e97e03a4a78e5ed995026a8c5549a445806c28cb03627\": rpc error: code = NotFound desc = could not find container \"457fa84c82e055fb574e97e03a4a78e5ed995026a8c5549a445806c28cb03627\": container with ID starting with 457fa84c82e055fb574e97e03a4a78e5ed995026a8c5549a445806c28cb03627 not found: ID does not exist" Dec 01 08:45:04 crc kubenswrapper[4873]: I1201 08:45:04.439122 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 01 08:45:05 crc kubenswrapper[4873]: I1201 08:45:05.429325 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:45:05 crc kubenswrapper[4873]: I1201 08:45:05.430423 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:45:05 crc kubenswrapper[4873]: I1201 08:45:05.646111 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-75494747d9-bhtkn"] Dec 01 08:45:06 crc kubenswrapper[4873]: I1201 08:45:06.444821 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:45:06 crc kubenswrapper[4873]: I1201 08:45:06.445382 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" event={"ID":"c624acad-0f23-48e9-95d1-3cffa69d3b04","Type":"ContainerStarted","Data":"b3d1c29e030f27c962354cbed0512418b10327fdf905be13164d847e0b3fcace"} Dec 01 08:45:06 crc kubenswrapper[4873]: I1201 08:45:06.445461 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" Dec 01 08:45:06 crc kubenswrapper[4873]: I1201 08:45:06.445484 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" event={"ID":"c624acad-0f23-48e9-95d1-3cffa69d3b04","Type":"ContainerStarted","Data":"18cd9166b6203abbd2aee4f517f36bedb7d702162ba7dad28eb84dee9f38e371"} Dec 01 08:45:06 crc kubenswrapper[4873]: I1201 08:45:06.462840 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-75494747d9-bhtkn" podStartSLOduration=66.462791191 podStartE2EDuration="1m6.462791191s" podCreationTimestamp="2025-12-01 08:44:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:45:06.459471913 +0000 UTC m=+282.361580492" watchObservedRunningTime="2025-12-01 08:45:06.462791191 +0000 UTC m=+282.364899770" Dec 01 08:45:22 crc kubenswrapper[4873]: I1201 08:45:22.542306 4873 generic.go:334] "Generic (PLEG): container finished" podID="2ea695f7-4e9e-4b12-8e0c-22431291576d" containerID="ac126d7386d526b4abbff206ecba9a518fe214fc8803ff9777184f9779560f9d" exitCode=0 Dec 01 08:45:22 crc kubenswrapper[4873]: I1201 08:45:22.542473 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7hstm" event={"ID":"2ea695f7-4e9e-4b12-8e0c-22431291576d","Type":"ContainerDied","Data":"ac126d7386d526b4abbff206ecba9a518fe214fc8803ff9777184f9779560f9d"} Dec 01 08:45:22 crc kubenswrapper[4873]: I1201 08:45:22.543921 4873 scope.go:117] "RemoveContainer" containerID="ac126d7386d526b4abbff206ecba9a518fe214fc8803ff9777184f9779560f9d" Dec 01 08:45:23 crc kubenswrapper[4873]: I1201 08:45:23.553373 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7hstm" event={"ID":"2ea695f7-4e9e-4b12-8e0c-22431291576d","Type":"ContainerStarted","Data":"358a739234fecdea83cf70ee42b928a350c890e435d33103c628751117eea789"} Dec 01 08:45:23 crc kubenswrapper[4873]: I1201 08:45:23.555424 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-7hstm" Dec 01 08:45:23 crc kubenswrapper[4873]: I1201 08:45:23.558115 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-7hstm" Dec 01 08:45:24 crc kubenswrapper[4873]: I1201 08:45:24.252884 4873 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Dec 01 08:45:26 crc kubenswrapper[4873]: I1201 08:45:26.367102 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-kw4m8"] Dec 01 08:45:26 crc kubenswrapper[4873]: I1201 08:45:26.367957 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-kw4m8" podUID="40a9c91a-1a6d-438d-9c65-51dc38b34f34" containerName="controller-manager" containerID="cri-o://e0c3f9fbfe354bd1a18ec89ef06c19691b818ad2eaac663b4713aebd080776a7" gracePeriod=30 Dec 01 08:45:26 crc kubenswrapper[4873]: I1201 08:45:26.472568 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-q4wmp"] Dec 01 08:45:26 crc kubenswrapper[4873]: I1201 08:45:26.472848 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q4wmp" podUID="83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5" containerName="route-controller-manager" containerID="cri-o://9124d6fa4df2506e2a75d9fe09d4a4aff028c3761c7506cfa5cbda7f6911b228" gracePeriod=30 Dec 01 08:45:26 crc kubenswrapper[4873]: I1201 08:45:26.577605 4873 generic.go:334] "Generic (PLEG): container finished" podID="40a9c91a-1a6d-438d-9c65-51dc38b34f34" containerID="e0c3f9fbfe354bd1a18ec89ef06c19691b818ad2eaac663b4713aebd080776a7" exitCode=0 Dec 01 08:45:26 crc kubenswrapper[4873]: I1201 08:45:26.577775 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-kw4m8" event={"ID":"40a9c91a-1a6d-438d-9c65-51dc38b34f34","Type":"ContainerDied","Data":"e0c3f9fbfe354bd1a18ec89ef06c19691b818ad2eaac663b4713aebd080776a7"} Dec 01 08:45:26 crc kubenswrapper[4873]: I1201 08:45:26.752752 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-kw4m8" Dec 01 08:45:26 crc kubenswrapper[4873]: I1201 08:45:26.875626 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nf47\" (UniqueName: \"kubernetes.io/projected/40a9c91a-1a6d-438d-9c65-51dc38b34f34-kube-api-access-8nf47\") pod \"40a9c91a-1a6d-438d-9c65-51dc38b34f34\" (UID: \"40a9c91a-1a6d-438d-9c65-51dc38b34f34\") " Dec 01 08:45:26 crc kubenswrapper[4873]: I1201 08:45:26.875674 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/40a9c91a-1a6d-438d-9c65-51dc38b34f34-proxy-ca-bundles\") pod \"40a9c91a-1a6d-438d-9c65-51dc38b34f34\" (UID: \"40a9c91a-1a6d-438d-9c65-51dc38b34f34\") " Dec 01 08:45:26 crc kubenswrapper[4873]: I1201 08:45:26.875704 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40a9c91a-1a6d-438d-9c65-51dc38b34f34-config\") pod \"40a9c91a-1a6d-438d-9c65-51dc38b34f34\" (UID: \"40a9c91a-1a6d-438d-9c65-51dc38b34f34\") " Dec 01 08:45:26 crc kubenswrapper[4873]: I1201 08:45:26.875725 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40a9c91a-1a6d-438d-9c65-51dc38b34f34-serving-cert\") pod \"40a9c91a-1a6d-438d-9c65-51dc38b34f34\" (UID: \"40a9c91a-1a6d-438d-9c65-51dc38b34f34\") " Dec 01 08:45:26 crc kubenswrapper[4873]: I1201 08:45:26.875758 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/40a9c91a-1a6d-438d-9c65-51dc38b34f34-client-ca\") pod \"40a9c91a-1a6d-438d-9c65-51dc38b34f34\" (UID: \"40a9c91a-1a6d-438d-9c65-51dc38b34f34\") " Dec 01 08:45:26 crc kubenswrapper[4873]: I1201 08:45:26.879756 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40a9c91a-1a6d-438d-9c65-51dc38b34f34-client-ca" (OuterVolumeSpecName: "client-ca") pod "40a9c91a-1a6d-438d-9c65-51dc38b34f34" (UID: "40a9c91a-1a6d-438d-9c65-51dc38b34f34"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:45:26 crc kubenswrapper[4873]: I1201 08:45:26.879898 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40a9c91a-1a6d-438d-9c65-51dc38b34f34-config" (OuterVolumeSpecName: "config") pod "40a9c91a-1a6d-438d-9c65-51dc38b34f34" (UID: "40a9c91a-1a6d-438d-9c65-51dc38b34f34"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:45:26 crc kubenswrapper[4873]: I1201 08:45:26.880591 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40a9c91a-1a6d-438d-9c65-51dc38b34f34-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "40a9c91a-1a6d-438d-9c65-51dc38b34f34" (UID: "40a9c91a-1a6d-438d-9c65-51dc38b34f34"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:45:26 crc kubenswrapper[4873]: I1201 08:45:26.898362 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40a9c91a-1a6d-438d-9c65-51dc38b34f34-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "40a9c91a-1a6d-438d-9c65-51dc38b34f34" (UID: "40a9c91a-1a6d-438d-9c65-51dc38b34f34"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:45:26 crc kubenswrapper[4873]: I1201 08:45:26.898471 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40a9c91a-1a6d-438d-9c65-51dc38b34f34-kube-api-access-8nf47" (OuterVolumeSpecName: "kube-api-access-8nf47") pod "40a9c91a-1a6d-438d-9c65-51dc38b34f34" (UID: "40a9c91a-1a6d-438d-9c65-51dc38b34f34"). InnerVolumeSpecName "kube-api-access-8nf47". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:45:26 crc kubenswrapper[4873]: I1201 08:45:26.924050 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q4wmp" Dec 01 08:45:26 crc kubenswrapper[4873]: I1201 08:45:26.977247 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8nf47\" (UniqueName: \"kubernetes.io/projected/40a9c91a-1a6d-438d-9c65-51dc38b34f34-kube-api-access-8nf47\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:26 crc kubenswrapper[4873]: I1201 08:45:26.977289 4873 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/40a9c91a-1a6d-438d-9c65-51dc38b34f34-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:26 crc kubenswrapper[4873]: I1201 08:45:26.977302 4873 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40a9c91a-1a6d-438d-9c65-51dc38b34f34-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:26 crc kubenswrapper[4873]: I1201 08:45:26.977312 4873 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40a9c91a-1a6d-438d-9c65-51dc38b34f34-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:26 crc kubenswrapper[4873]: I1201 08:45:26.977323 4873 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/40a9c91a-1a6d-438d-9c65-51dc38b34f34-client-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:27 crc kubenswrapper[4873]: I1201 08:45:27.078849 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5-client-ca\") pod \"83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5\" (UID: \"83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5\") " Dec 01 08:45:27 crc kubenswrapper[4873]: I1201 08:45:27.078920 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5-serving-cert\") pod \"83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5\" (UID: \"83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5\") " Dec 01 08:45:27 crc kubenswrapper[4873]: I1201 08:45:27.078973 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qwvk\" (UniqueName: \"kubernetes.io/projected/83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5-kube-api-access-4qwvk\") pod \"83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5\" (UID: \"83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5\") " Dec 01 08:45:27 crc kubenswrapper[4873]: I1201 08:45:27.079051 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5-config\") pod \"83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5\" (UID: \"83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5\") " Dec 01 08:45:27 crc kubenswrapper[4873]: I1201 08:45:27.080123 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5-client-ca" (OuterVolumeSpecName: "client-ca") pod "83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5" (UID: "83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:45:27 crc kubenswrapper[4873]: I1201 08:45:27.080148 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5-config" (OuterVolumeSpecName: "config") pod "83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5" (UID: "83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:45:27 crc kubenswrapper[4873]: I1201 08:45:27.083802 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5" (UID: "83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:45:27 crc kubenswrapper[4873]: I1201 08:45:27.085735 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5-kube-api-access-4qwvk" (OuterVolumeSpecName: "kube-api-access-4qwvk") pod "83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5" (UID: "83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5"). InnerVolumeSpecName "kube-api-access-4qwvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:45:27 crc kubenswrapper[4873]: I1201 08:45:27.181425 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qwvk\" (UniqueName: \"kubernetes.io/projected/83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5-kube-api-access-4qwvk\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:27 crc kubenswrapper[4873]: I1201 08:45:27.181506 4873 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:27 crc kubenswrapper[4873]: I1201 08:45:27.181532 4873 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5-client-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:27 crc kubenswrapper[4873]: I1201 08:45:27.181557 4873 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:27 crc kubenswrapper[4873]: I1201 08:45:27.584684 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-kw4m8" event={"ID":"40a9c91a-1a6d-438d-9c65-51dc38b34f34","Type":"ContainerDied","Data":"80c19d1e61c087c6b2acfbee6dd4fbbbe839dd042bde472e13444d35c71f2c0a"} Dec 01 08:45:27 crc kubenswrapper[4873]: I1201 08:45:27.585181 4873 scope.go:117] "RemoveContainer" containerID="e0c3f9fbfe354bd1a18ec89ef06c19691b818ad2eaac663b4713aebd080776a7" Dec 01 08:45:27 crc kubenswrapper[4873]: I1201 08:45:27.585325 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-kw4m8" Dec 01 08:45:27 crc kubenswrapper[4873]: I1201 08:45:27.591034 4873 generic.go:334] "Generic (PLEG): container finished" podID="83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5" containerID="9124d6fa4df2506e2a75d9fe09d4a4aff028c3761c7506cfa5cbda7f6911b228" exitCode=0 Dec 01 08:45:27 crc kubenswrapper[4873]: I1201 08:45:27.591080 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q4wmp" event={"ID":"83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5","Type":"ContainerDied","Data":"9124d6fa4df2506e2a75d9fe09d4a4aff028c3761c7506cfa5cbda7f6911b228"} Dec 01 08:45:27 crc kubenswrapper[4873]: I1201 08:45:27.591094 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q4wmp" Dec 01 08:45:27 crc kubenswrapper[4873]: I1201 08:45:27.591110 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q4wmp" event={"ID":"83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5","Type":"ContainerDied","Data":"cd5b64af3eea625553caa17357eda8565a7cef89341a8b134d32d289d053e807"} Dec 01 08:45:27 crc kubenswrapper[4873]: I1201 08:45:27.613010 4873 scope.go:117] "RemoveContainer" containerID="9124d6fa4df2506e2a75d9fe09d4a4aff028c3761c7506cfa5cbda7f6911b228" Dec 01 08:45:27 crc kubenswrapper[4873]: I1201 08:45:27.623332 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-q4wmp"] Dec 01 08:45:27 crc kubenswrapper[4873]: I1201 08:45:27.630990 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-q4wmp"] Dec 01 08:45:27 crc kubenswrapper[4873]: I1201 08:45:27.637077 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-kw4m8"] Dec 01 08:45:27 crc kubenswrapper[4873]: I1201 08:45:27.642207 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-kw4m8"] Dec 01 08:45:27 crc kubenswrapper[4873]: I1201 08:45:27.647217 4873 scope.go:117] "RemoveContainer" containerID="9124d6fa4df2506e2a75d9fe09d4a4aff028c3761c7506cfa5cbda7f6911b228" Dec 01 08:45:27 crc kubenswrapper[4873]: E1201 08:45:27.647964 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9124d6fa4df2506e2a75d9fe09d4a4aff028c3761c7506cfa5cbda7f6911b228\": container with ID starting with 9124d6fa4df2506e2a75d9fe09d4a4aff028c3761c7506cfa5cbda7f6911b228 not found: ID does not exist" containerID="9124d6fa4df2506e2a75d9fe09d4a4aff028c3761c7506cfa5cbda7f6911b228" Dec 01 08:45:27 crc kubenswrapper[4873]: I1201 08:45:27.648039 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9124d6fa4df2506e2a75d9fe09d4a4aff028c3761c7506cfa5cbda7f6911b228"} err="failed to get container status \"9124d6fa4df2506e2a75d9fe09d4a4aff028c3761c7506cfa5cbda7f6911b228\": rpc error: code = NotFound desc = could not find container \"9124d6fa4df2506e2a75d9fe09d4a4aff028c3761c7506cfa5cbda7f6911b228\": container with ID starting with 9124d6fa4df2506e2a75d9fe09d4a4aff028c3761c7506cfa5cbda7f6911b228 not found: ID does not exist" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.440483 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40a9c91a-1a6d-438d-9c65-51dc38b34f34" path="/var/lib/kubelet/pods/40a9c91a-1a6d-438d-9c65-51dc38b34f34/volumes" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.442081 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5" path="/var/lib/kubelet/pods/83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5/volumes" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.492400 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-569854d77f-cjtfx"] Dec 01 08:45:28 crc kubenswrapper[4873]: E1201 08:45:28.492707 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40a9c91a-1a6d-438d-9c65-51dc38b34f34" containerName="controller-manager" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.492730 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="40a9c91a-1a6d-438d-9c65-51dc38b34f34" containerName="controller-manager" Dec 01 08:45:28 crc kubenswrapper[4873]: E1201 08:45:28.492740 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6d9174b-c4b0-4176-9e95-291c04608323" containerName="collect-profiles" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.492747 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6d9174b-c4b0-4176-9e95-291c04608323" containerName="collect-profiles" Dec 01 08:45:28 crc kubenswrapper[4873]: E1201 08:45:28.492765 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5" containerName="route-controller-manager" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.492773 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5" containerName="route-controller-manager" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.492896 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="83dd20fa-39b3-4e5c-ae3c-b06674f6f2c5" containerName="route-controller-manager" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.492913 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="40a9c91a-1a6d-438d-9c65-51dc38b34f34" containerName="controller-manager" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.492925 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6d9174b-c4b0-4176-9e95-291c04608323" containerName="collect-profiles" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.493372 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-569854d77f-cjtfx" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.497184 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.497229 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66649d5fd7-zcb7m"] Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.497564 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.497913 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-66649d5fd7-zcb7m" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.499780 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.499912 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.500154 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.500716 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.501077 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66649d5fd7-zcb7m"] Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.502088 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.502188 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.502258 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.502318 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.502372 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.502383 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.506441 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-569854d77f-cjtfx"] Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.507991 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.605215 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prgkh\" (UniqueName: \"kubernetes.io/projected/bde4cec0-f82d-4724-9a01-f41a2bb3c93e-kube-api-access-prgkh\") pod \"controller-manager-569854d77f-cjtfx\" (UID: \"bde4cec0-f82d-4724-9a01-f41a2bb3c93e\") " pod="openshift-controller-manager/controller-manager-569854d77f-cjtfx" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.605305 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bde4cec0-f82d-4724-9a01-f41a2bb3c93e-serving-cert\") pod \"controller-manager-569854d77f-cjtfx\" (UID: \"bde4cec0-f82d-4724-9a01-f41a2bb3c93e\") " pod="openshift-controller-manager/controller-manager-569854d77f-cjtfx" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.605534 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdvqv\" (UniqueName: \"kubernetes.io/projected/a6608aa5-ae02-444c-a9a7-b50ee7fa2bee-kube-api-access-rdvqv\") pod \"route-controller-manager-66649d5fd7-zcb7m\" (UID: \"a6608aa5-ae02-444c-a9a7-b50ee7fa2bee\") " pod="openshift-route-controller-manager/route-controller-manager-66649d5fd7-zcb7m" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.605627 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6608aa5-ae02-444c-a9a7-b50ee7fa2bee-config\") pod \"route-controller-manager-66649d5fd7-zcb7m\" (UID: \"a6608aa5-ae02-444c-a9a7-b50ee7fa2bee\") " pod="openshift-route-controller-manager/route-controller-manager-66649d5fd7-zcb7m" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.605659 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a6608aa5-ae02-444c-a9a7-b50ee7fa2bee-serving-cert\") pod \"route-controller-manager-66649d5fd7-zcb7m\" (UID: \"a6608aa5-ae02-444c-a9a7-b50ee7fa2bee\") " pod="openshift-route-controller-manager/route-controller-manager-66649d5fd7-zcb7m" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.605712 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bde4cec0-f82d-4724-9a01-f41a2bb3c93e-client-ca\") pod \"controller-manager-569854d77f-cjtfx\" (UID: \"bde4cec0-f82d-4724-9a01-f41a2bb3c93e\") " pod="openshift-controller-manager/controller-manager-569854d77f-cjtfx" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.605735 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a6608aa5-ae02-444c-a9a7-b50ee7fa2bee-client-ca\") pod \"route-controller-manager-66649d5fd7-zcb7m\" (UID: \"a6608aa5-ae02-444c-a9a7-b50ee7fa2bee\") " pod="openshift-route-controller-manager/route-controller-manager-66649d5fd7-zcb7m" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.605880 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bde4cec0-f82d-4724-9a01-f41a2bb3c93e-proxy-ca-bundles\") pod \"controller-manager-569854d77f-cjtfx\" (UID: \"bde4cec0-f82d-4724-9a01-f41a2bb3c93e\") " pod="openshift-controller-manager/controller-manager-569854d77f-cjtfx" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.606076 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bde4cec0-f82d-4724-9a01-f41a2bb3c93e-config\") pod \"controller-manager-569854d77f-cjtfx\" (UID: \"bde4cec0-f82d-4724-9a01-f41a2bb3c93e\") " pod="openshift-controller-manager/controller-manager-569854d77f-cjtfx" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.707715 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bde4cec0-f82d-4724-9a01-f41a2bb3c93e-config\") pod \"controller-manager-569854d77f-cjtfx\" (UID: \"bde4cec0-f82d-4724-9a01-f41a2bb3c93e\") " pod="openshift-controller-manager/controller-manager-569854d77f-cjtfx" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.707810 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prgkh\" (UniqueName: \"kubernetes.io/projected/bde4cec0-f82d-4724-9a01-f41a2bb3c93e-kube-api-access-prgkh\") pod \"controller-manager-569854d77f-cjtfx\" (UID: \"bde4cec0-f82d-4724-9a01-f41a2bb3c93e\") " pod="openshift-controller-manager/controller-manager-569854d77f-cjtfx" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.707843 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bde4cec0-f82d-4724-9a01-f41a2bb3c93e-serving-cert\") pod \"controller-manager-569854d77f-cjtfx\" (UID: \"bde4cec0-f82d-4724-9a01-f41a2bb3c93e\") " pod="openshift-controller-manager/controller-manager-569854d77f-cjtfx" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.707903 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdvqv\" (UniqueName: \"kubernetes.io/projected/a6608aa5-ae02-444c-a9a7-b50ee7fa2bee-kube-api-access-rdvqv\") pod \"route-controller-manager-66649d5fd7-zcb7m\" (UID: \"a6608aa5-ae02-444c-a9a7-b50ee7fa2bee\") " pod="openshift-route-controller-manager/route-controller-manager-66649d5fd7-zcb7m" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.707931 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6608aa5-ae02-444c-a9a7-b50ee7fa2bee-config\") pod \"route-controller-manager-66649d5fd7-zcb7m\" (UID: \"a6608aa5-ae02-444c-a9a7-b50ee7fa2bee\") " pod="openshift-route-controller-manager/route-controller-manager-66649d5fd7-zcb7m" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.707953 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a6608aa5-ae02-444c-a9a7-b50ee7fa2bee-serving-cert\") pod \"route-controller-manager-66649d5fd7-zcb7m\" (UID: \"a6608aa5-ae02-444c-a9a7-b50ee7fa2bee\") " pod="openshift-route-controller-manager/route-controller-manager-66649d5fd7-zcb7m" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.707980 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bde4cec0-f82d-4724-9a01-f41a2bb3c93e-client-ca\") pod \"controller-manager-569854d77f-cjtfx\" (UID: \"bde4cec0-f82d-4724-9a01-f41a2bb3c93e\") " pod="openshift-controller-manager/controller-manager-569854d77f-cjtfx" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.708000 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a6608aa5-ae02-444c-a9a7-b50ee7fa2bee-client-ca\") pod \"route-controller-manager-66649d5fd7-zcb7m\" (UID: \"a6608aa5-ae02-444c-a9a7-b50ee7fa2bee\") " pod="openshift-route-controller-manager/route-controller-manager-66649d5fd7-zcb7m" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.708043 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bde4cec0-f82d-4724-9a01-f41a2bb3c93e-proxy-ca-bundles\") pod \"controller-manager-569854d77f-cjtfx\" (UID: \"bde4cec0-f82d-4724-9a01-f41a2bb3c93e\") " pod="openshift-controller-manager/controller-manager-569854d77f-cjtfx" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.710006 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bde4cec0-f82d-4724-9a01-f41a2bb3c93e-client-ca\") pod \"controller-manager-569854d77f-cjtfx\" (UID: \"bde4cec0-f82d-4724-9a01-f41a2bb3c93e\") " pod="openshift-controller-manager/controller-manager-569854d77f-cjtfx" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.710104 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6608aa5-ae02-444c-a9a7-b50ee7fa2bee-config\") pod \"route-controller-manager-66649d5fd7-zcb7m\" (UID: \"a6608aa5-ae02-444c-a9a7-b50ee7fa2bee\") " pod="openshift-route-controller-manager/route-controller-manager-66649d5fd7-zcb7m" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.710185 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bde4cec0-f82d-4724-9a01-f41a2bb3c93e-config\") pod \"controller-manager-569854d77f-cjtfx\" (UID: \"bde4cec0-f82d-4724-9a01-f41a2bb3c93e\") " pod="openshift-controller-manager/controller-manager-569854d77f-cjtfx" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.710279 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bde4cec0-f82d-4724-9a01-f41a2bb3c93e-proxy-ca-bundles\") pod \"controller-manager-569854d77f-cjtfx\" (UID: \"bde4cec0-f82d-4724-9a01-f41a2bb3c93e\") " pod="openshift-controller-manager/controller-manager-569854d77f-cjtfx" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.710950 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a6608aa5-ae02-444c-a9a7-b50ee7fa2bee-client-ca\") pod \"route-controller-manager-66649d5fd7-zcb7m\" (UID: \"a6608aa5-ae02-444c-a9a7-b50ee7fa2bee\") " pod="openshift-route-controller-manager/route-controller-manager-66649d5fd7-zcb7m" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.717055 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a6608aa5-ae02-444c-a9a7-b50ee7fa2bee-serving-cert\") pod \"route-controller-manager-66649d5fd7-zcb7m\" (UID: \"a6608aa5-ae02-444c-a9a7-b50ee7fa2bee\") " pod="openshift-route-controller-manager/route-controller-manager-66649d5fd7-zcb7m" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.719542 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bde4cec0-f82d-4724-9a01-f41a2bb3c93e-serving-cert\") pod \"controller-manager-569854d77f-cjtfx\" (UID: \"bde4cec0-f82d-4724-9a01-f41a2bb3c93e\") " pod="openshift-controller-manager/controller-manager-569854d77f-cjtfx" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.732654 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prgkh\" (UniqueName: \"kubernetes.io/projected/bde4cec0-f82d-4724-9a01-f41a2bb3c93e-kube-api-access-prgkh\") pod \"controller-manager-569854d77f-cjtfx\" (UID: \"bde4cec0-f82d-4724-9a01-f41a2bb3c93e\") " pod="openshift-controller-manager/controller-manager-569854d77f-cjtfx" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.736671 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdvqv\" (UniqueName: \"kubernetes.io/projected/a6608aa5-ae02-444c-a9a7-b50ee7fa2bee-kube-api-access-rdvqv\") pod \"route-controller-manager-66649d5fd7-zcb7m\" (UID: \"a6608aa5-ae02-444c-a9a7-b50ee7fa2bee\") " pod="openshift-route-controller-manager/route-controller-manager-66649d5fd7-zcb7m" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.814667 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-569854d77f-cjtfx" Dec 01 08:45:28 crc kubenswrapper[4873]: I1201 08:45:28.830159 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-66649d5fd7-zcb7m" Dec 01 08:45:29 crc kubenswrapper[4873]: I1201 08:45:29.040841 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66649d5fd7-zcb7m"] Dec 01 08:45:29 crc kubenswrapper[4873]: I1201 08:45:29.086666 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-569854d77f-cjtfx"] Dec 01 08:45:29 crc kubenswrapper[4873]: W1201 08:45:29.096525 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbde4cec0_f82d_4724_9a01_f41a2bb3c93e.slice/crio-ab777310183fb70f88c457eb88e2a9cf56afefca07659854bc242432d0f149fc WatchSource:0}: Error finding container ab777310183fb70f88c457eb88e2a9cf56afefca07659854bc242432d0f149fc: Status 404 returned error can't find the container with id ab777310183fb70f88c457eb88e2a9cf56afefca07659854bc242432d0f149fc Dec 01 08:45:29 crc kubenswrapper[4873]: I1201 08:45:29.608743 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-569854d77f-cjtfx" event={"ID":"bde4cec0-f82d-4724-9a01-f41a2bb3c93e","Type":"ContainerStarted","Data":"e503baf6a0d2d429f2b3b50847b9e38384b54a49a1ef5b092f5d6e14c1c76f65"} Dec 01 08:45:29 crc kubenswrapper[4873]: I1201 08:45:29.609300 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-569854d77f-cjtfx" event={"ID":"bde4cec0-f82d-4724-9a01-f41a2bb3c93e","Type":"ContainerStarted","Data":"ab777310183fb70f88c457eb88e2a9cf56afefca07659854bc242432d0f149fc"} Dec 01 08:45:29 crc kubenswrapper[4873]: I1201 08:45:29.609803 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-569854d77f-cjtfx" Dec 01 08:45:29 crc kubenswrapper[4873]: I1201 08:45:29.613305 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-66649d5fd7-zcb7m" event={"ID":"a6608aa5-ae02-444c-a9a7-b50ee7fa2bee","Type":"ContainerStarted","Data":"faba54e86835a830f98929a8715e5f332c3e65db89554f46dde5af720cac5b97"} Dec 01 08:45:29 crc kubenswrapper[4873]: I1201 08:45:29.613339 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-66649d5fd7-zcb7m" event={"ID":"a6608aa5-ae02-444c-a9a7-b50ee7fa2bee","Type":"ContainerStarted","Data":"763e98f8a9bb25c1c72a66d258aa1fcc43dc66b48dcd2adde7335927a3904b65"} Dec 01 08:45:29 crc kubenswrapper[4873]: I1201 08:45:29.613928 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-66649d5fd7-zcb7m" Dec 01 08:45:29 crc kubenswrapper[4873]: I1201 08:45:29.618479 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-569854d77f-cjtfx" Dec 01 08:45:29 crc kubenswrapper[4873]: I1201 08:45:29.634488 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-569854d77f-cjtfx" podStartSLOduration=3.634464772 podStartE2EDuration="3.634464772s" podCreationTimestamp="2025-12-01 08:45:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:45:29.633965581 +0000 UTC m=+305.536074120" watchObservedRunningTime="2025-12-01 08:45:29.634464772 +0000 UTC m=+305.536573311" Dec 01 08:45:29 crc kubenswrapper[4873]: I1201 08:45:29.698712 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-66649d5fd7-zcb7m" podStartSLOduration=3.698673681 podStartE2EDuration="3.698673681s" podCreationTimestamp="2025-12-01 08:45:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:45:29.68375315 +0000 UTC m=+305.585861689" watchObservedRunningTime="2025-12-01 08:45:29.698673681 +0000 UTC m=+305.600782220" Dec 01 08:45:29 crc kubenswrapper[4873]: I1201 08:45:29.920764 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-66649d5fd7-zcb7m" Dec 01 08:46:01 crc kubenswrapper[4873]: I1201 08:46:01.058995 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:46:01 crc kubenswrapper[4873]: I1201 08:46:01.059856 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.185229 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jr8nt"] Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.187624 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jr8nt" podUID="d44d4877-74e9-4fdc-a062-92d031f1636d" containerName="registry-server" containerID="cri-o://24ec8dd1d9493b9bda7288ab158dd7d3dfe73a1525a5fb6e20b5770a7a49f8f8" gracePeriod=30 Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.218812 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dm6vl"] Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.219268 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-dm6vl" podUID="7a18563f-c79c-4fed-83fb-5ee7d865e014" containerName="registry-server" containerID="cri-o://0e4fe44a1413ee54a8747489eee92e455858698931e5b2a4f1dfc41c98e70158" gracePeriod=30 Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.236110 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7hstm"] Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.236551 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-7hstm" podUID="2ea695f7-4e9e-4b12-8e0c-22431291576d" containerName="marketplace-operator" containerID="cri-o://358a739234fecdea83cf70ee42b928a350c890e435d33103c628751117eea789" gracePeriod=30 Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.237141 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-spckc"] Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.237433 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-spckc" podUID="73df40c6-5524-496f-92ba-36100c4af8bb" containerName="registry-server" containerID="cri-o://0bb44c93f1cbd0aa1641ba7202566d8d91f2e2a3d3180ac1e2e0897f44fd2fd1" gracePeriod=30 Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.250766 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zffmt"] Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.251097 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zffmt" podUID="a3fc6e9a-1447-41e2-8886-e5ea32e3d353" containerName="registry-server" containerID="cri-o://f9aaaf1883f656b600c4e791acc271b8a0214606fb549554c4ff0b1a5a3f3977" gracePeriod=30 Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.263030 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dgdcc"] Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.264228 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-dgdcc" Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.268320 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dgdcc"] Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.403572 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brw4w\" (UniqueName: \"kubernetes.io/projected/c0e7b6b5-8852-4dec-bbf2-b7247b34a791-kube-api-access-brw4w\") pod \"marketplace-operator-79b997595-dgdcc\" (UID: \"c0e7b6b5-8852-4dec-bbf2-b7247b34a791\") " pod="openshift-marketplace/marketplace-operator-79b997595-dgdcc" Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.403640 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c0e7b6b5-8852-4dec-bbf2-b7247b34a791-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-dgdcc\" (UID: \"c0e7b6b5-8852-4dec-bbf2-b7247b34a791\") " pod="openshift-marketplace/marketplace-operator-79b997595-dgdcc" Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.403672 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c0e7b6b5-8852-4dec-bbf2-b7247b34a791-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-dgdcc\" (UID: \"c0e7b6b5-8852-4dec-bbf2-b7247b34a791\") " pod="openshift-marketplace/marketplace-operator-79b997595-dgdcc" Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.505383 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brw4w\" (UniqueName: \"kubernetes.io/projected/c0e7b6b5-8852-4dec-bbf2-b7247b34a791-kube-api-access-brw4w\") pod \"marketplace-operator-79b997595-dgdcc\" (UID: \"c0e7b6b5-8852-4dec-bbf2-b7247b34a791\") " pod="openshift-marketplace/marketplace-operator-79b997595-dgdcc" Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.505433 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c0e7b6b5-8852-4dec-bbf2-b7247b34a791-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-dgdcc\" (UID: \"c0e7b6b5-8852-4dec-bbf2-b7247b34a791\") " pod="openshift-marketplace/marketplace-operator-79b997595-dgdcc" Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.505863 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c0e7b6b5-8852-4dec-bbf2-b7247b34a791-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-dgdcc\" (UID: \"c0e7b6b5-8852-4dec-bbf2-b7247b34a791\") " pod="openshift-marketplace/marketplace-operator-79b997595-dgdcc" Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.509096 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c0e7b6b5-8852-4dec-bbf2-b7247b34a791-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-dgdcc\" (UID: \"c0e7b6b5-8852-4dec-bbf2-b7247b34a791\") " pod="openshift-marketplace/marketplace-operator-79b997595-dgdcc" Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.513337 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c0e7b6b5-8852-4dec-bbf2-b7247b34a791-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-dgdcc\" (UID: \"c0e7b6b5-8852-4dec-bbf2-b7247b34a791\") " pod="openshift-marketplace/marketplace-operator-79b997595-dgdcc" Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.521882 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brw4w\" (UniqueName: \"kubernetes.io/projected/c0e7b6b5-8852-4dec-bbf2-b7247b34a791-kube-api-access-brw4w\") pod \"marketplace-operator-79b997595-dgdcc\" (UID: \"c0e7b6b5-8852-4dec-bbf2-b7247b34a791\") " pod="openshift-marketplace/marketplace-operator-79b997595-dgdcc" Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.579918 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-dgdcc" Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.683398 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jr8nt" Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.810779 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gg4z6\" (UniqueName: \"kubernetes.io/projected/d44d4877-74e9-4fdc-a062-92d031f1636d-kube-api-access-gg4z6\") pod \"d44d4877-74e9-4fdc-a062-92d031f1636d\" (UID: \"d44d4877-74e9-4fdc-a062-92d031f1636d\") " Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.810937 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d44d4877-74e9-4fdc-a062-92d031f1636d-utilities\") pod \"d44d4877-74e9-4fdc-a062-92d031f1636d\" (UID: \"d44d4877-74e9-4fdc-a062-92d031f1636d\") " Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.811007 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d44d4877-74e9-4fdc-a062-92d031f1636d-catalog-content\") pod \"d44d4877-74e9-4fdc-a062-92d031f1636d\" (UID: \"d44d4877-74e9-4fdc-a062-92d031f1636d\") " Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.812836 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d44d4877-74e9-4fdc-a062-92d031f1636d-utilities" (OuterVolumeSpecName: "utilities") pod "d44d4877-74e9-4fdc-a062-92d031f1636d" (UID: "d44d4877-74e9-4fdc-a062-92d031f1636d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.817490 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d44d4877-74e9-4fdc-a062-92d031f1636d-kube-api-access-gg4z6" (OuterVolumeSpecName: "kube-api-access-gg4z6") pod "d44d4877-74e9-4fdc-a062-92d031f1636d" (UID: "d44d4877-74e9-4fdc-a062-92d031f1636d"). InnerVolumeSpecName "kube-api-access-gg4z6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.844865 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7hstm" Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.854209 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-spckc" Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.861337 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zffmt" Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.872762 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dm6vl" Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.880458 4873 generic.go:334] "Generic (PLEG): container finished" podID="73df40c6-5524-496f-92ba-36100c4af8bb" containerID="0bb44c93f1cbd0aa1641ba7202566d8d91f2e2a3d3180ac1e2e0897f44fd2fd1" exitCode=0 Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.880605 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-spckc" event={"ID":"73df40c6-5524-496f-92ba-36100c4af8bb","Type":"ContainerDied","Data":"0bb44c93f1cbd0aa1641ba7202566d8d91f2e2a3d3180ac1e2e0897f44fd2fd1"} Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.880648 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-spckc" event={"ID":"73df40c6-5524-496f-92ba-36100c4af8bb","Type":"ContainerDied","Data":"e2c9ca91af2103d1045c7daf00e5177e3c887669f99479ad1ca9b091eabc6446"} Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.880709 4873 scope.go:117] "RemoveContainer" containerID="0bb44c93f1cbd0aa1641ba7202566d8d91f2e2a3d3180ac1e2e0897f44fd2fd1" Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.884209 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-spckc" Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.895428 4873 generic.go:334] "Generic (PLEG): container finished" podID="2ea695f7-4e9e-4b12-8e0c-22431291576d" containerID="358a739234fecdea83cf70ee42b928a350c890e435d33103c628751117eea789" exitCode=0 Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.895605 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7hstm" Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.896051 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7hstm" event={"ID":"2ea695f7-4e9e-4b12-8e0c-22431291576d","Type":"ContainerDied","Data":"358a739234fecdea83cf70ee42b928a350c890e435d33103c628751117eea789"} Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.896089 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7hstm" event={"ID":"2ea695f7-4e9e-4b12-8e0c-22431291576d","Type":"ContainerDied","Data":"4813d2fb668949029e06c337a682b5bbc38ebdf66e908a6ff62a80f644365233"} Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.907346 4873 generic.go:334] "Generic (PLEG): container finished" podID="7a18563f-c79c-4fed-83fb-5ee7d865e014" containerID="0e4fe44a1413ee54a8747489eee92e455858698931e5b2a4f1dfc41c98e70158" exitCode=0 Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.907439 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dm6vl" event={"ID":"7a18563f-c79c-4fed-83fb-5ee7d865e014","Type":"ContainerDied","Data":"0e4fe44a1413ee54a8747489eee92e455858698931e5b2a4f1dfc41c98e70158"} Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.907477 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dm6vl" event={"ID":"7a18563f-c79c-4fed-83fb-5ee7d865e014","Type":"ContainerDied","Data":"811c5560bd438367a21c284150b9c460ab798fe6be675bdc68322faf6c604744"} Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.907558 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dm6vl" Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.913733 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gg4z6\" (UniqueName: \"kubernetes.io/projected/d44d4877-74e9-4fdc-a062-92d031f1636d-kube-api-access-gg4z6\") on node \"crc\" DevicePath \"\"" Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.913784 4873 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d44d4877-74e9-4fdc-a062-92d031f1636d-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.919269 4873 scope.go:117] "RemoveContainer" containerID="6a87de807ce1c794686eccb6fc12010fb7df329698853264ae3cb744cf34effe" Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.923627 4873 generic.go:334] "Generic (PLEG): container finished" podID="a3fc6e9a-1447-41e2-8886-e5ea32e3d353" containerID="f9aaaf1883f656b600c4e791acc271b8a0214606fb549554c4ff0b1a5a3f3977" exitCode=0 Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.923721 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zffmt" event={"ID":"a3fc6e9a-1447-41e2-8886-e5ea32e3d353","Type":"ContainerDied","Data":"f9aaaf1883f656b600c4e791acc271b8a0214606fb549554c4ff0b1a5a3f3977"} Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.923762 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zffmt" event={"ID":"a3fc6e9a-1447-41e2-8886-e5ea32e3d353","Type":"ContainerDied","Data":"f803672e16cf9ec4069e2f400930ee609bf4f10cae3d61260a212f8ef760cc9b"} Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.923842 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zffmt" Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.933390 4873 generic.go:334] "Generic (PLEG): container finished" podID="d44d4877-74e9-4fdc-a062-92d031f1636d" containerID="24ec8dd1d9493b9bda7288ab158dd7d3dfe73a1525a5fb6e20b5770a7a49f8f8" exitCode=0 Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.933435 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jr8nt" event={"ID":"d44d4877-74e9-4fdc-a062-92d031f1636d","Type":"ContainerDied","Data":"24ec8dd1d9493b9bda7288ab158dd7d3dfe73a1525a5fb6e20b5770a7a49f8f8"} Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.933463 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jr8nt" event={"ID":"d44d4877-74e9-4fdc-a062-92d031f1636d","Type":"ContainerDied","Data":"0f0ac101db567bf9028ba04cc6e1fb6fd0d75b2b1e6ac471cd2df8eb0ebf7b9f"} Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.933615 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jr8nt" Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.939763 4873 scope.go:117] "RemoveContainer" containerID="c7133b53533ce8651130c3291d66c8d102fb954b3e2ccf5164c777a7182e66fb" Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.939753 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d44d4877-74e9-4fdc-a062-92d031f1636d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d44d4877-74e9-4fdc-a062-92d031f1636d" (UID: "d44d4877-74e9-4fdc-a062-92d031f1636d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.958754 4873 scope.go:117] "RemoveContainer" containerID="0bb44c93f1cbd0aa1641ba7202566d8d91f2e2a3d3180ac1e2e0897f44fd2fd1" Dec 01 08:46:04 crc kubenswrapper[4873]: E1201 08:46:04.960889 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0bb44c93f1cbd0aa1641ba7202566d8d91f2e2a3d3180ac1e2e0897f44fd2fd1\": container with ID starting with 0bb44c93f1cbd0aa1641ba7202566d8d91f2e2a3d3180ac1e2e0897f44fd2fd1 not found: ID does not exist" containerID="0bb44c93f1cbd0aa1641ba7202566d8d91f2e2a3d3180ac1e2e0897f44fd2fd1" Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.960937 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bb44c93f1cbd0aa1641ba7202566d8d91f2e2a3d3180ac1e2e0897f44fd2fd1"} err="failed to get container status \"0bb44c93f1cbd0aa1641ba7202566d8d91f2e2a3d3180ac1e2e0897f44fd2fd1\": rpc error: code = NotFound desc = could not find container \"0bb44c93f1cbd0aa1641ba7202566d8d91f2e2a3d3180ac1e2e0897f44fd2fd1\": container with ID starting with 0bb44c93f1cbd0aa1641ba7202566d8d91f2e2a3d3180ac1e2e0897f44fd2fd1 not found: ID does not exist" Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.960970 4873 scope.go:117] "RemoveContainer" containerID="6a87de807ce1c794686eccb6fc12010fb7df329698853264ae3cb744cf34effe" Dec 01 08:46:04 crc kubenswrapper[4873]: E1201 08:46:04.961370 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a87de807ce1c794686eccb6fc12010fb7df329698853264ae3cb744cf34effe\": container with ID starting with 6a87de807ce1c794686eccb6fc12010fb7df329698853264ae3cb744cf34effe not found: ID does not exist" containerID="6a87de807ce1c794686eccb6fc12010fb7df329698853264ae3cb744cf34effe" Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.961388 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a87de807ce1c794686eccb6fc12010fb7df329698853264ae3cb744cf34effe"} err="failed to get container status \"6a87de807ce1c794686eccb6fc12010fb7df329698853264ae3cb744cf34effe\": rpc error: code = NotFound desc = could not find container \"6a87de807ce1c794686eccb6fc12010fb7df329698853264ae3cb744cf34effe\": container with ID starting with 6a87de807ce1c794686eccb6fc12010fb7df329698853264ae3cb744cf34effe not found: ID does not exist" Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.961399 4873 scope.go:117] "RemoveContainer" containerID="c7133b53533ce8651130c3291d66c8d102fb954b3e2ccf5164c777a7182e66fb" Dec 01 08:46:04 crc kubenswrapper[4873]: E1201 08:46:04.961579 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7133b53533ce8651130c3291d66c8d102fb954b3e2ccf5164c777a7182e66fb\": container with ID starting with c7133b53533ce8651130c3291d66c8d102fb954b3e2ccf5164c777a7182e66fb not found: ID does not exist" containerID="c7133b53533ce8651130c3291d66c8d102fb954b3e2ccf5164c777a7182e66fb" Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.961598 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7133b53533ce8651130c3291d66c8d102fb954b3e2ccf5164c777a7182e66fb"} err="failed to get container status \"c7133b53533ce8651130c3291d66c8d102fb954b3e2ccf5164c777a7182e66fb\": rpc error: code = NotFound desc = could not find container \"c7133b53533ce8651130c3291d66c8d102fb954b3e2ccf5164c777a7182e66fb\": container with ID starting with c7133b53533ce8651130c3291d66c8d102fb954b3e2ccf5164c777a7182e66fb not found: ID does not exist" Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.961611 4873 scope.go:117] "RemoveContainer" containerID="358a739234fecdea83cf70ee42b928a350c890e435d33103c628751117eea789" Dec 01 08:46:04 crc kubenswrapper[4873]: I1201 08:46:04.987725 4873 scope.go:117] "RemoveContainer" containerID="ac126d7386d526b4abbff206ecba9a518fe214fc8803ff9777184f9779560f9d" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.005781 4873 scope.go:117] "RemoveContainer" containerID="358a739234fecdea83cf70ee42b928a350c890e435d33103c628751117eea789" Dec 01 08:46:05 crc kubenswrapper[4873]: E1201 08:46:05.006779 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"358a739234fecdea83cf70ee42b928a350c890e435d33103c628751117eea789\": container with ID starting with 358a739234fecdea83cf70ee42b928a350c890e435d33103c628751117eea789 not found: ID does not exist" containerID="358a739234fecdea83cf70ee42b928a350c890e435d33103c628751117eea789" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.006829 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"358a739234fecdea83cf70ee42b928a350c890e435d33103c628751117eea789"} err="failed to get container status \"358a739234fecdea83cf70ee42b928a350c890e435d33103c628751117eea789\": rpc error: code = NotFound desc = could not find container \"358a739234fecdea83cf70ee42b928a350c890e435d33103c628751117eea789\": container with ID starting with 358a739234fecdea83cf70ee42b928a350c890e435d33103c628751117eea789 not found: ID does not exist" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.006873 4873 scope.go:117] "RemoveContainer" containerID="ac126d7386d526b4abbff206ecba9a518fe214fc8803ff9777184f9779560f9d" Dec 01 08:46:05 crc kubenswrapper[4873]: E1201 08:46:05.007253 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac126d7386d526b4abbff206ecba9a518fe214fc8803ff9777184f9779560f9d\": container with ID starting with ac126d7386d526b4abbff206ecba9a518fe214fc8803ff9777184f9779560f9d not found: ID does not exist" containerID="ac126d7386d526b4abbff206ecba9a518fe214fc8803ff9777184f9779560f9d" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.007316 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac126d7386d526b4abbff206ecba9a518fe214fc8803ff9777184f9779560f9d"} err="failed to get container status \"ac126d7386d526b4abbff206ecba9a518fe214fc8803ff9777184f9779560f9d\": rpc error: code = NotFound desc = could not find container \"ac126d7386d526b4abbff206ecba9a518fe214fc8803ff9777184f9779560f9d\": container with ID starting with ac126d7386d526b4abbff206ecba9a518fe214fc8803ff9777184f9779560f9d not found: ID does not exist" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.007393 4873 scope.go:117] "RemoveContainer" containerID="0e4fe44a1413ee54a8747489eee92e455858698931e5b2a4f1dfc41c98e70158" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.015134 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-flxnb\" (UniqueName: \"kubernetes.io/projected/73df40c6-5524-496f-92ba-36100c4af8bb-kube-api-access-flxnb\") pod \"73df40c6-5524-496f-92ba-36100c4af8bb\" (UID: \"73df40c6-5524-496f-92ba-36100c4af8bb\") " Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.015189 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73df40c6-5524-496f-92ba-36100c4af8bb-catalog-content\") pod \"73df40c6-5524-496f-92ba-36100c4af8bb\" (UID: \"73df40c6-5524-496f-92ba-36100c4af8bb\") " Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.015222 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pl7wx\" (UniqueName: \"kubernetes.io/projected/7a18563f-c79c-4fed-83fb-5ee7d865e014-kube-api-access-pl7wx\") pod \"7a18563f-c79c-4fed-83fb-5ee7d865e014\" (UID: \"7a18563f-c79c-4fed-83fb-5ee7d865e014\") " Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.015277 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b8s74\" (UniqueName: \"kubernetes.io/projected/a3fc6e9a-1447-41e2-8886-e5ea32e3d353-kube-api-access-b8s74\") pod \"a3fc6e9a-1447-41e2-8886-e5ea32e3d353\" (UID: \"a3fc6e9a-1447-41e2-8886-e5ea32e3d353\") " Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.015313 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3fc6e9a-1447-41e2-8886-e5ea32e3d353-utilities\") pod \"a3fc6e9a-1447-41e2-8886-e5ea32e3d353\" (UID: \"a3fc6e9a-1447-41e2-8886-e5ea32e3d353\") " Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.015397 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a18563f-c79c-4fed-83fb-5ee7d865e014-utilities\") pod \"7a18563f-c79c-4fed-83fb-5ee7d865e014\" (UID: \"7a18563f-c79c-4fed-83fb-5ee7d865e014\") " Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.015423 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2ea695f7-4e9e-4b12-8e0c-22431291576d-marketplace-trusted-ca\") pod \"2ea695f7-4e9e-4b12-8e0c-22431291576d\" (UID: \"2ea695f7-4e9e-4b12-8e0c-22431291576d\") " Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.015454 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73df40c6-5524-496f-92ba-36100c4af8bb-utilities\") pod \"73df40c6-5524-496f-92ba-36100c4af8bb\" (UID: \"73df40c6-5524-496f-92ba-36100c4af8bb\") " Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.015480 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2ea695f7-4e9e-4b12-8e0c-22431291576d-marketplace-operator-metrics\") pod \"2ea695f7-4e9e-4b12-8e0c-22431291576d\" (UID: \"2ea695f7-4e9e-4b12-8e0c-22431291576d\") " Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.015540 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3fc6e9a-1447-41e2-8886-e5ea32e3d353-catalog-content\") pod \"a3fc6e9a-1447-41e2-8886-e5ea32e3d353\" (UID: \"a3fc6e9a-1447-41e2-8886-e5ea32e3d353\") " Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.015567 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a18563f-c79c-4fed-83fb-5ee7d865e014-catalog-content\") pod \"7a18563f-c79c-4fed-83fb-5ee7d865e014\" (UID: \"7a18563f-c79c-4fed-83fb-5ee7d865e014\") " Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.015593 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-72t8c\" (UniqueName: \"kubernetes.io/projected/2ea695f7-4e9e-4b12-8e0c-22431291576d-kube-api-access-72t8c\") pod \"2ea695f7-4e9e-4b12-8e0c-22431291576d\" (UID: \"2ea695f7-4e9e-4b12-8e0c-22431291576d\") " Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.015837 4873 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d44d4877-74e9-4fdc-a062-92d031f1636d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.016523 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a18563f-c79c-4fed-83fb-5ee7d865e014-utilities" (OuterVolumeSpecName: "utilities") pod "7a18563f-c79c-4fed-83fb-5ee7d865e014" (UID: "7a18563f-c79c-4fed-83fb-5ee7d865e014"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.016629 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3fc6e9a-1447-41e2-8886-e5ea32e3d353-utilities" (OuterVolumeSpecName: "utilities") pod "a3fc6e9a-1447-41e2-8886-e5ea32e3d353" (UID: "a3fc6e9a-1447-41e2-8886-e5ea32e3d353"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.016648 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ea695f7-4e9e-4b12-8e0c-22431291576d-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "2ea695f7-4e9e-4b12-8e0c-22431291576d" (UID: "2ea695f7-4e9e-4b12-8e0c-22431291576d"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.016856 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73df40c6-5524-496f-92ba-36100c4af8bb-utilities" (OuterVolumeSpecName: "utilities") pod "73df40c6-5524-496f-92ba-36100c4af8bb" (UID: "73df40c6-5524-496f-92ba-36100c4af8bb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.020841 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ea695f7-4e9e-4b12-8e0c-22431291576d-kube-api-access-72t8c" (OuterVolumeSpecName: "kube-api-access-72t8c") pod "2ea695f7-4e9e-4b12-8e0c-22431291576d" (UID: "2ea695f7-4e9e-4b12-8e0c-22431291576d"). InnerVolumeSpecName "kube-api-access-72t8c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.020900 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73df40c6-5524-496f-92ba-36100c4af8bb-kube-api-access-flxnb" (OuterVolumeSpecName: "kube-api-access-flxnb") pod "73df40c6-5524-496f-92ba-36100c4af8bb" (UID: "73df40c6-5524-496f-92ba-36100c4af8bb"). InnerVolumeSpecName "kube-api-access-flxnb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.021115 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ea695f7-4e9e-4b12-8e0c-22431291576d-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "2ea695f7-4e9e-4b12-8e0c-22431291576d" (UID: "2ea695f7-4e9e-4b12-8e0c-22431291576d"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.022058 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a18563f-c79c-4fed-83fb-5ee7d865e014-kube-api-access-pl7wx" (OuterVolumeSpecName: "kube-api-access-pl7wx") pod "7a18563f-c79c-4fed-83fb-5ee7d865e014" (UID: "7a18563f-c79c-4fed-83fb-5ee7d865e014"). InnerVolumeSpecName "kube-api-access-pl7wx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.029831 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3fc6e9a-1447-41e2-8886-e5ea32e3d353-kube-api-access-b8s74" (OuterVolumeSpecName: "kube-api-access-b8s74") pod "a3fc6e9a-1447-41e2-8886-e5ea32e3d353" (UID: "a3fc6e9a-1447-41e2-8886-e5ea32e3d353"). InnerVolumeSpecName "kube-api-access-b8s74". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.031672 4873 scope.go:117] "RemoveContainer" containerID="2c7e82be231e5ba969e6f77cb86c791d7aeda2b15bc412839ead5ad6caeddf82" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.052093 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73df40c6-5524-496f-92ba-36100c4af8bb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "73df40c6-5524-496f-92ba-36100c4af8bb" (UID: "73df40c6-5524-496f-92ba-36100c4af8bb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.053155 4873 scope.go:117] "RemoveContainer" containerID="66fcda06001cbdaccd70eea2d66862bc764bf489bf2ffe5da590e6223e8109dd" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.080867 4873 scope.go:117] "RemoveContainer" containerID="0e4fe44a1413ee54a8747489eee92e455858698931e5b2a4f1dfc41c98e70158" Dec 01 08:46:05 crc kubenswrapper[4873]: E1201 08:46:05.081703 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e4fe44a1413ee54a8747489eee92e455858698931e5b2a4f1dfc41c98e70158\": container with ID starting with 0e4fe44a1413ee54a8747489eee92e455858698931e5b2a4f1dfc41c98e70158 not found: ID does not exist" containerID="0e4fe44a1413ee54a8747489eee92e455858698931e5b2a4f1dfc41c98e70158" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.081808 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e4fe44a1413ee54a8747489eee92e455858698931e5b2a4f1dfc41c98e70158"} err="failed to get container status \"0e4fe44a1413ee54a8747489eee92e455858698931e5b2a4f1dfc41c98e70158\": rpc error: code = NotFound desc = could not find container \"0e4fe44a1413ee54a8747489eee92e455858698931e5b2a4f1dfc41c98e70158\": container with ID starting with 0e4fe44a1413ee54a8747489eee92e455858698931e5b2a4f1dfc41c98e70158 not found: ID does not exist" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.081849 4873 scope.go:117] "RemoveContainer" containerID="2c7e82be231e5ba969e6f77cb86c791d7aeda2b15bc412839ead5ad6caeddf82" Dec 01 08:46:05 crc kubenswrapper[4873]: E1201 08:46:05.082253 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c7e82be231e5ba969e6f77cb86c791d7aeda2b15bc412839ead5ad6caeddf82\": container with ID starting with 2c7e82be231e5ba969e6f77cb86c791d7aeda2b15bc412839ead5ad6caeddf82 not found: ID does not exist" containerID="2c7e82be231e5ba969e6f77cb86c791d7aeda2b15bc412839ead5ad6caeddf82" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.082300 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c7e82be231e5ba969e6f77cb86c791d7aeda2b15bc412839ead5ad6caeddf82"} err="failed to get container status \"2c7e82be231e5ba969e6f77cb86c791d7aeda2b15bc412839ead5ad6caeddf82\": rpc error: code = NotFound desc = could not find container \"2c7e82be231e5ba969e6f77cb86c791d7aeda2b15bc412839ead5ad6caeddf82\": container with ID starting with 2c7e82be231e5ba969e6f77cb86c791d7aeda2b15bc412839ead5ad6caeddf82 not found: ID does not exist" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.082318 4873 scope.go:117] "RemoveContainer" containerID="66fcda06001cbdaccd70eea2d66862bc764bf489bf2ffe5da590e6223e8109dd" Dec 01 08:46:05 crc kubenswrapper[4873]: E1201 08:46:05.082573 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66fcda06001cbdaccd70eea2d66862bc764bf489bf2ffe5da590e6223e8109dd\": container with ID starting with 66fcda06001cbdaccd70eea2d66862bc764bf489bf2ffe5da590e6223e8109dd not found: ID does not exist" containerID="66fcda06001cbdaccd70eea2d66862bc764bf489bf2ffe5da590e6223e8109dd" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.082613 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66fcda06001cbdaccd70eea2d66862bc764bf489bf2ffe5da590e6223e8109dd"} err="failed to get container status \"66fcda06001cbdaccd70eea2d66862bc764bf489bf2ffe5da590e6223e8109dd\": rpc error: code = NotFound desc = could not find container \"66fcda06001cbdaccd70eea2d66862bc764bf489bf2ffe5da590e6223e8109dd\": container with ID starting with 66fcda06001cbdaccd70eea2d66862bc764bf489bf2ffe5da590e6223e8109dd not found: ID does not exist" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.082629 4873 scope.go:117] "RemoveContainer" containerID="f9aaaf1883f656b600c4e791acc271b8a0214606fb549554c4ff0b1a5a3f3977" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.090193 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dgdcc"] Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.090276 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a18563f-c79c-4fed-83fb-5ee7d865e014-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7a18563f-c79c-4fed-83fb-5ee7d865e014" (UID: "7a18563f-c79c-4fed-83fb-5ee7d865e014"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.100427 4873 scope.go:117] "RemoveContainer" containerID="895095f340c0d6b6c537639212d41c0b567a44775d21e76ba90c62ffe4bd3ef8" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.117045 4873 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a18563f-c79c-4fed-83fb-5ee7d865e014-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.117074 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-72t8c\" (UniqueName: \"kubernetes.io/projected/2ea695f7-4e9e-4b12-8e0c-22431291576d-kube-api-access-72t8c\") on node \"crc\" DevicePath \"\"" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.117090 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-flxnb\" (UniqueName: \"kubernetes.io/projected/73df40c6-5524-496f-92ba-36100c4af8bb-kube-api-access-flxnb\") on node \"crc\" DevicePath \"\"" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.117100 4873 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73df40c6-5524-496f-92ba-36100c4af8bb-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.117110 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pl7wx\" (UniqueName: \"kubernetes.io/projected/7a18563f-c79c-4fed-83fb-5ee7d865e014-kube-api-access-pl7wx\") on node \"crc\" DevicePath \"\"" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.117121 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b8s74\" (UniqueName: \"kubernetes.io/projected/a3fc6e9a-1447-41e2-8886-e5ea32e3d353-kube-api-access-b8s74\") on node \"crc\" DevicePath \"\"" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.117131 4873 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3fc6e9a-1447-41e2-8886-e5ea32e3d353-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.117140 4873 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a18563f-c79c-4fed-83fb-5ee7d865e014-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.117153 4873 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2ea695f7-4e9e-4b12-8e0c-22431291576d-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.117164 4873 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73df40c6-5524-496f-92ba-36100c4af8bb-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.117172 4873 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2ea695f7-4e9e-4b12-8e0c-22431291576d-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.123090 4873 scope.go:117] "RemoveContainer" containerID="e030f35f012ce934c7de7c1f6faee445f1761fb194016df182d3be267a7a0c88" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.139713 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3fc6e9a-1447-41e2-8886-e5ea32e3d353-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a3fc6e9a-1447-41e2-8886-e5ea32e3d353" (UID: "a3fc6e9a-1447-41e2-8886-e5ea32e3d353"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.141391 4873 scope.go:117] "RemoveContainer" containerID="f9aaaf1883f656b600c4e791acc271b8a0214606fb549554c4ff0b1a5a3f3977" Dec 01 08:46:05 crc kubenswrapper[4873]: E1201 08:46:05.141791 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9aaaf1883f656b600c4e791acc271b8a0214606fb549554c4ff0b1a5a3f3977\": container with ID starting with f9aaaf1883f656b600c4e791acc271b8a0214606fb549554c4ff0b1a5a3f3977 not found: ID does not exist" containerID="f9aaaf1883f656b600c4e791acc271b8a0214606fb549554c4ff0b1a5a3f3977" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.141831 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9aaaf1883f656b600c4e791acc271b8a0214606fb549554c4ff0b1a5a3f3977"} err="failed to get container status \"f9aaaf1883f656b600c4e791acc271b8a0214606fb549554c4ff0b1a5a3f3977\": rpc error: code = NotFound desc = could not find container \"f9aaaf1883f656b600c4e791acc271b8a0214606fb549554c4ff0b1a5a3f3977\": container with ID starting with f9aaaf1883f656b600c4e791acc271b8a0214606fb549554c4ff0b1a5a3f3977 not found: ID does not exist" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.141861 4873 scope.go:117] "RemoveContainer" containerID="895095f340c0d6b6c537639212d41c0b567a44775d21e76ba90c62ffe4bd3ef8" Dec 01 08:46:05 crc kubenswrapper[4873]: E1201 08:46:05.142114 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"895095f340c0d6b6c537639212d41c0b567a44775d21e76ba90c62ffe4bd3ef8\": container with ID starting with 895095f340c0d6b6c537639212d41c0b567a44775d21e76ba90c62ffe4bd3ef8 not found: ID does not exist" containerID="895095f340c0d6b6c537639212d41c0b567a44775d21e76ba90c62ffe4bd3ef8" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.142148 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"895095f340c0d6b6c537639212d41c0b567a44775d21e76ba90c62ffe4bd3ef8"} err="failed to get container status \"895095f340c0d6b6c537639212d41c0b567a44775d21e76ba90c62ffe4bd3ef8\": rpc error: code = NotFound desc = could not find container \"895095f340c0d6b6c537639212d41c0b567a44775d21e76ba90c62ffe4bd3ef8\": container with ID starting with 895095f340c0d6b6c537639212d41c0b567a44775d21e76ba90c62ffe4bd3ef8 not found: ID does not exist" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.142166 4873 scope.go:117] "RemoveContainer" containerID="e030f35f012ce934c7de7c1f6faee445f1761fb194016df182d3be267a7a0c88" Dec 01 08:46:05 crc kubenswrapper[4873]: E1201 08:46:05.144416 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e030f35f012ce934c7de7c1f6faee445f1761fb194016df182d3be267a7a0c88\": container with ID starting with e030f35f012ce934c7de7c1f6faee445f1761fb194016df182d3be267a7a0c88 not found: ID does not exist" containerID="e030f35f012ce934c7de7c1f6faee445f1761fb194016df182d3be267a7a0c88" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.144495 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e030f35f012ce934c7de7c1f6faee445f1761fb194016df182d3be267a7a0c88"} err="failed to get container status \"e030f35f012ce934c7de7c1f6faee445f1761fb194016df182d3be267a7a0c88\": rpc error: code = NotFound desc = could not find container \"e030f35f012ce934c7de7c1f6faee445f1761fb194016df182d3be267a7a0c88\": container with ID starting with e030f35f012ce934c7de7c1f6faee445f1761fb194016df182d3be267a7a0c88 not found: ID does not exist" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.144535 4873 scope.go:117] "RemoveContainer" containerID="24ec8dd1d9493b9bda7288ab158dd7d3dfe73a1525a5fb6e20b5770a7a49f8f8" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.166500 4873 scope.go:117] "RemoveContainer" containerID="4d52f6e6e3e9a7b6e7d24181236c9333a0e82333e1da5b5d4f9fa17f228e47ff" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.215614 4873 scope.go:117] "RemoveContainer" containerID="76bf7b52eeff1ac99c60219084b867f2cd1abb47d6fea23a99461dea51d606a5" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.223073 4873 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3fc6e9a-1447-41e2-8886-e5ea32e3d353-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.242924 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-spckc"] Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.247122 4873 scope.go:117] "RemoveContainer" containerID="24ec8dd1d9493b9bda7288ab158dd7d3dfe73a1525a5fb6e20b5770a7a49f8f8" Dec 01 08:46:05 crc kubenswrapper[4873]: E1201 08:46:05.248100 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24ec8dd1d9493b9bda7288ab158dd7d3dfe73a1525a5fb6e20b5770a7a49f8f8\": container with ID starting with 24ec8dd1d9493b9bda7288ab158dd7d3dfe73a1525a5fb6e20b5770a7a49f8f8 not found: ID does not exist" containerID="24ec8dd1d9493b9bda7288ab158dd7d3dfe73a1525a5fb6e20b5770a7a49f8f8" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.248143 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24ec8dd1d9493b9bda7288ab158dd7d3dfe73a1525a5fb6e20b5770a7a49f8f8"} err="failed to get container status \"24ec8dd1d9493b9bda7288ab158dd7d3dfe73a1525a5fb6e20b5770a7a49f8f8\": rpc error: code = NotFound desc = could not find container \"24ec8dd1d9493b9bda7288ab158dd7d3dfe73a1525a5fb6e20b5770a7a49f8f8\": container with ID starting with 24ec8dd1d9493b9bda7288ab158dd7d3dfe73a1525a5fb6e20b5770a7a49f8f8 not found: ID does not exist" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.248174 4873 scope.go:117] "RemoveContainer" containerID="4d52f6e6e3e9a7b6e7d24181236c9333a0e82333e1da5b5d4f9fa17f228e47ff" Dec 01 08:46:05 crc kubenswrapper[4873]: E1201 08:46:05.248546 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d52f6e6e3e9a7b6e7d24181236c9333a0e82333e1da5b5d4f9fa17f228e47ff\": container with ID starting with 4d52f6e6e3e9a7b6e7d24181236c9333a0e82333e1da5b5d4f9fa17f228e47ff not found: ID does not exist" containerID="4d52f6e6e3e9a7b6e7d24181236c9333a0e82333e1da5b5d4f9fa17f228e47ff" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.248580 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d52f6e6e3e9a7b6e7d24181236c9333a0e82333e1da5b5d4f9fa17f228e47ff"} err="failed to get container status \"4d52f6e6e3e9a7b6e7d24181236c9333a0e82333e1da5b5d4f9fa17f228e47ff\": rpc error: code = NotFound desc = could not find container \"4d52f6e6e3e9a7b6e7d24181236c9333a0e82333e1da5b5d4f9fa17f228e47ff\": container with ID starting with 4d52f6e6e3e9a7b6e7d24181236c9333a0e82333e1da5b5d4f9fa17f228e47ff not found: ID does not exist" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.248603 4873 scope.go:117] "RemoveContainer" containerID="76bf7b52eeff1ac99c60219084b867f2cd1abb47d6fea23a99461dea51d606a5" Dec 01 08:46:05 crc kubenswrapper[4873]: E1201 08:46:05.249069 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76bf7b52eeff1ac99c60219084b867f2cd1abb47d6fea23a99461dea51d606a5\": container with ID starting with 76bf7b52eeff1ac99c60219084b867f2cd1abb47d6fea23a99461dea51d606a5 not found: ID does not exist" containerID="76bf7b52eeff1ac99c60219084b867f2cd1abb47d6fea23a99461dea51d606a5" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.249123 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76bf7b52eeff1ac99c60219084b867f2cd1abb47d6fea23a99461dea51d606a5"} err="failed to get container status \"76bf7b52eeff1ac99c60219084b867f2cd1abb47d6fea23a99461dea51d606a5\": rpc error: code = NotFound desc = could not find container \"76bf7b52eeff1ac99c60219084b867f2cd1abb47d6fea23a99461dea51d606a5\": container with ID starting with 76bf7b52eeff1ac99c60219084b867f2cd1abb47d6fea23a99461dea51d606a5 not found: ID does not exist" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.257790 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-spckc"] Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.262371 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7hstm"] Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.268121 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7hstm"] Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.273760 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dm6vl"] Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.280384 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-dm6vl"] Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.286829 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jr8nt"] Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.295042 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jr8nt"] Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.305898 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zffmt"] Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.310694 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zffmt"] Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.950879 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-dgdcc" event={"ID":"c0e7b6b5-8852-4dec-bbf2-b7247b34a791","Type":"ContainerStarted","Data":"56fc505fca6ffb546076632cbdc95373f8b0aa32dbb202d051205d933a34ac75"} Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.951494 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-dgdcc" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.951511 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-dgdcc" event={"ID":"c0e7b6b5-8852-4dec-bbf2-b7247b34a791","Type":"ContainerStarted","Data":"d7fc6adcb104db5347c440fe0d7b89f834f3dd0fb78f9fefea8c13f0c34faf22"} Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.954871 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-dgdcc" Dec 01 08:46:05 crc kubenswrapper[4873]: I1201 08:46:05.977183 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-dgdcc" podStartSLOduration=1.977160133 podStartE2EDuration="1.977160133s" podCreationTimestamp="2025-12-01 08:46:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:46:05.975655367 +0000 UTC m=+341.877763916" watchObservedRunningTime="2025-12-01 08:46:05.977160133 +0000 UTC m=+341.879268682" Dec 01 08:46:06 crc kubenswrapper[4873]: I1201 08:46:06.355090 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-569854d77f-cjtfx"] Dec 01 08:46:06 crc kubenswrapper[4873]: I1201 08:46:06.355536 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-569854d77f-cjtfx" podUID="bde4cec0-f82d-4724-9a01-f41a2bb3c93e" containerName="controller-manager" containerID="cri-o://e503baf6a0d2d429f2b3b50847b9e38384b54a49a1ef5b092f5d6e14c1c76f65" gracePeriod=30 Dec 01 08:46:06 crc kubenswrapper[4873]: I1201 08:46:06.436434 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ea695f7-4e9e-4b12-8e0c-22431291576d" path="/var/lib/kubelet/pods/2ea695f7-4e9e-4b12-8e0c-22431291576d/volumes" Dec 01 08:46:06 crc kubenswrapper[4873]: I1201 08:46:06.436970 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73df40c6-5524-496f-92ba-36100c4af8bb" path="/var/lib/kubelet/pods/73df40c6-5524-496f-92ba-36100c4af8bb/volumes" Dec 01 08:46:06 crc kubenswrapper[4873]: I1201 08:46:06.437635 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a18563f-c79c-4fed-83fb-5ee7d865e014" path="/var/lib/kubelet/pods/7a18563f-c79c-4fed-83fb-5ee7d865e014/volumes" Dec 01 08:46:06 crc kubenswrapper[4873]: I1201 08:46:06.438270 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3fc6e9a-1447-41e2-8886-e5ea32e3d353" path="/var/lib/kubelet/pods/a3fc6e9a-1447-41e2-8886-e5ea32e3d353/volumes" Dec 01 08:46:06 crc kubenswrapper[4873]: I1201 08:46:06.438904 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d44d4877-74e9-4fdc-a062-92d031f1636d" path="/var/lib/kubelet/pods/d44d4877-74e9-4fdc-a062-92d031f1636d/volumes" Dec 01 08:46:06 crc kubenswrapper[4873]: I1201 08:46:06.774934 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-569854d77f-cjtfx" Dec 01 08:46:06 crc kubenswrapper[4873]: I1201 08:46:06.948955 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bde4cec0-f82d-4724-9a01-f41a2bb3c93e-config\") pod \"bde4cec0-f82d-4724-9a01-f41a2bb3c93e\" (UID: \"bde4cec0-f82d-4724-9a01-f41a2bb3c93e\") " Dec 01 08:46:06 crc kubenswrapper[4873]: I1201 08:46:06.949739 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bde4cec0-f82d-4724-9a01-f41a2bb3c93e-client-ca\") pod \"bde4cec0-f82d-4724-9a01-f41a2bb3c93e\" (UID: \"bde4cec0-f82d-4724-9a01-f41a2bb3c93e\") " Dec 01 08:46:06 crc kubenswrapper[4873]: I1201 08:46:06.949804 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bde4cec0-f82d-4724-9a01-f41a2bb3c93e-serving-cert\") pod \"bde4cec0-f82d-4724-9a01-f41a2bb3c93e\" (UID: \"bde4cec0-f82d-4724-9a01-f41a2bb3c93e\") " Dec 01 08:46:06 crc kubenswrapper[4873]: I1201 08:46:06.949829 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bde4cec0-f82d-4724-9a01-f41a2bb3c93e-proxy-ca-bundles\") pod \"bde4cec0-f82d-4724-9a01-f41a2bb3c93e\" (UID: \"bde4cec0-f82d-4724-9a01-f41a2bb3c93e\") " Dec 01 08:46:06 crc kubenswrapper[4873]: I1201 08:46:06.949888 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-prgkh\" (UniqueName: \"kubernetes.io/projected/bde4cec0-f82d-4724-9a01-f41a2bb3c93e-kube-api-access-prgkh\") pod \"bde4cec0-f82d-4724-9a01-f41a2bb3c93e\" (UID: \"bde4cec0-f82d-4724-9a01-f41a2bb3c93e\") " Dec 01 08:46:06 crc kubenswrapper[4873]: I1201 08:46:06.950260 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bde4cec0-f82d-4724-9a01-f41a2bb3c93e-config" (OuterVolumeSpecName: "config") pod "bde4cec0-f82d-4724-9a01-f41a2bb3c93e" (UID: "bde4cec0-f82d-4724-9a01-f41a2bb3c93e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:46:06 crc kubenswrapper[4873]: I1201 08:46:06.950951 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bde4cec0-f82d-4724-9a01-f41a2bb3c93e-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "bde4cec0-f82d-4724-9a01-f41a2bb3c93e" (UID: "bde4cec0-f82d-4724-9a01-f41a2bb3c93e"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:46:06 crc kubenswrapper[4873]: I1201 08:46:06.951038 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bde4cec0-f82d-4724-9a01-f41a2bb3c93e-client-ca" (OuterVolumeSpecName: "client-ca") pod "bde4cec0-f82d-4724-9a01-f41a2bb3c93e" (UID: "bde4cec0-f82d-4724-9a01-f41a2bb3c93e"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:46:06 crc kubenswrapper[4873]: I1201 08:46:06.959433 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bde4cec0-f82d-4724-9a01-f41a2bb3c93e-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bde4cec0-f82d-4724-9a01-f41a2bb3c93e" (UID: "bde4cec0-f82d-4724-9a01-f41a2bb3c93e"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:46:06 crc kubenswrapper[4873]: I1201 08:46:06.959733 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bde4cec0-f82d-4724-9a01-f41a2bb3c93e-kube-api-access-prgkh" (OuterVolumeSpecName: "kube-api-access-prgkh") pod "bde4cec0-f82d-4724-9a01-f41a2bb3c93e" (UID: "bde4cec0-f82d-4724-9a01-f41a2bb3c93e"). InnerVolumeSpecName "kube-api-access-prgkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:46:06 crc kubenswrapper[4873]: I1201 08:46:06.974295 4873 generic.go:334] "Generic (PLEG): container finished" podID="bde4cec0-f82d-4724-9a01-f41a2bb3c93e" containerID="e503baf6a0d2d429f2b3b50847b9e38384b54a49a1ef5b092f5d6e14c1c76f65" exitCode=0 Dec 01 08:46:06 crc kubenswrapper[4873]: I1201 08:46:06.974415 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-569854d77f-cjtfx" Dec 01 08:46:06 crc kubenswrapper[4873]: I1201 08:46:06.974604 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-569854d77f-cjtfx" event={"ID":"bde4cec0-f82d-4724-9a01-f41a2bb3c93e","Type":"ContainerDied","Data":"e503baf6a0d2d429f2b3b50847b9e38384b54a49a1ef5b092f5d6e14c1c76f65"} Dec 01 08:46:06 crc kubenswrapper[4873]: I1201 08:46:06.974751 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-569854d77f-cjtfx" event={"ID":"bde4cec0-f82d-4724-9a01-f41a2bb3c93e","Type":"ContainerDied","Data":"ab777310183fb70f88c457eb88e2a9cf56afefca07659854bc242432d0f149fc"} Dec 01 08:46:06 crc kubenswrapper[4873]: I1201 08:46:06.974848 4873 scope.go:117] "RemoveContainer" containerID="e503baf6a0d2d429f2b3b50847b9e38384b54a49a1ef5b092f5d6e14c1c76f65" Dec 01 08:46:06 crc kubenswrapper[4873]: I1201 08:46:06.999086 4873 scope.go:117] "RemoveContainer" containerID="e503baf6a0d2d429f2b3b50847b9e38384b54a49a1ef5b092f5d6e14c1c76f65" Dec 01 08:46:06 crc kubenswrapper[4873]: E1201 08:46:06.999763 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e503baf6a0d2d429f2b3b50847b9e38384b54a49a1ef5b092f5d6e14c1c76f65\": container with ID starting with e503baf6a0d2d429f2b3b50847b9e38384b54a49a1ef5b092f5d6e14c1c76f65 not found: ID does not exist" containerID="e503baf6a0d2d429f2b3b50847b9e38384b54a49a1ef5b092f5d6e14c1c76f65" Dec 01 08:46:06 crc kubenswrapper[4873]: I1201 08:46:06.999818 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e503baf6a0d2d429f2b3b50847b9e38384b54a49a1ef5b092f5d6e14c1c76f65"} err="failed to get container status \"e503baf6a0d2d429f2b3b50847b9e38384b54a49a1ef5b092f5d6e14c1c76f65\": rpc error: code = NotFound desc = could not find container \"e503baf6a0d2d429f2b3b50847b9e38384b54a49a1ef5b092f5d6e14c1c76f65\": container with ID starting with e503baf6a0d2d429f2b3b50847b9e38384b54a49a1ef5b092f5d6e14c1c76f65 not found: ID does not exist" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.015534 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-569854d77f-cjtfx"] Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.018175 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-569854d77f-cjtfx"] Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.051930 4873 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bde4cec0-f82d-4724-9a01-f41a2bb3c93e-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.051989 4873 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bde4cec0-f82d-4724-9a01-f41a2bb3c93e-client-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.052002 4873 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bde4cec0-f82d-4724-9a01-f41a2bb3c93e-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.052010 4873 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bde4cec0-f82d-4724-9a01-f41a2bb3c93e-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.052032 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-prgkh\" (UniqueName: \"kubernetes.io/projected/bde4cec0-f82d-4724-9a01-f41a2bb3c93e-kube-api-access-prgkh\") on node \"crc\" DevicePath \"\"" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.516061 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7d57f65db7-hqchq"] Dec 01 08:46:07 crc kubenswrapper[4873]: E1201 08:46:07.516476 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a18563f-c79c-4fed-83fb-5ee7d865e014" containerName="registry-server" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.516519 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a18563f-c79c-4fed-83fb-5ee7d865e014" containerName="registry-server" Dec 01 08:46:07 crc kubenswrapper[4873]: E1201 08:46:07.516537 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bde4cec0-f82d-4724-9a01-f41a2bb3c93e" containerName="controller-manager" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.516544 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="bde4cec0-f82d-4724-9a01-f41a2bb3c93e" containerName="controller-manager" Dec 01 08:46:07 crc kubenswrapper[4873]: E1201 08:46:07.516553 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d44d4877-74e9-4fdc-a062-92d031f1636d" containerName="extract-utilities" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.516560 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="d44d4877-74e9-4fdc-a062-92d031f1636d" containerName="extract-utilities" Dec 01 08:46:07 crc kubenswrapper[4873]: E1201 08:46:07.516568 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3fc6e9a-1447-41e2-8886-e5ea32e3d353" containerName="registry-server" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.516576 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3fc6e9a-1447-41e2-8886-e5ea32e3d353" containerName="registry-server" Dec 01 08:46:07 crc kubenswrapper[4873]: E1201 08:46:07.516582 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ea695f7-4e9e-4b12-8e0c-22431291576d" containerName="marketplace-operator" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.516589 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ea695f7-4e9e-4b12-8e0c-22431291576d" containerName="marketplace-operator" Dec 01 08:46:07 crc kubenswrapper[4873]: E1201 08:46:07.516598 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73df40c6-5524-496f-92ba-36100c4af8bb" containerName="extract-content" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.516605 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="73df40c6-5524-496f-92ba-36100c4af8bb" containerName="extract-content" Dec 01 08:46:07 crc kubenswrapper[4873]: E1201 08:46:07.516614 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a18563f-c79c-4fed-83fb-5ee7d865e014" containerName="extract-utilities" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.516621 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a18563f-c79c-4fed-83fb-5ee7d865e014" containerName="extract-utilities" Dec 01 08:46:07 crc kubenswrapper[4873]: E1201 08:46:07.516631 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73df40c6-5524-496f-92ba-36100c4af8bb" containerName="registry-server" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.516637 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="73df40c6-5524-496f-92ba-36100c4af8bb" containerName="registry-server" Dec 01 08:46:07 crc kubenswrapper[4873]: E1201 08:46:07.516648 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d44d4877-74e9-4fdc-a062-92d031f1636d" containerName="registry-server" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.516655 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="d44d4877-74e9-4fdc-a062-92d031f1636d" containerName="registry-server" Dec 01 08:46:07 crc kubenswrapper[4873]: E1201 08:46:07.516670 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3fc6e9a-1447-41e2-8886-e5ea32e3d353" containerName="extract-content" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.516676 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3fc6e9a-1447-41e2-8886-e5ea32e3d353" containerName="extract-content" Dec 01 08:46:07 crc kubenswrapper[4873]: E1201 08:46:07.516688 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73df40c6-5524-496f-92ba-36100c4af8bb" containerName="extract-utilities" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.516694 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="73df40c6-5524-496f-92ba-36100c4af8bb" containerName="extract-utilities" Dec 01 08:46:07 crc kubenswrapper[4873]: E1201 08:46:07.516703 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ea695f7-4e9e-4b12-8e0c-22431291576d" containerName="marketplace-operator" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.516711 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ea695f7-4e9e-4b12-8e0c-22431291576d" containerName="marketplace-operator" Dec 01 08:46:07 crc kubenswrapper[4873]: E1201 08:46:07.516722 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d44d4877-74e9-4fdc-a062-92d031f1636d" containerName="extract-content" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.516729 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="d44d4877-74e9-4fdc-a062-92d031f1636d" containerName="extract-content" Dec 01 08:46:07 crc kubenswrapper[4873]: E1201 08:46:07.516740 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3fc6e9a-1447-41e2-8886-e5ea32e3d353" containerName="extract-utilities" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.516748 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3fc6e9a-1447-41e2-8886-e5ea32e3d353" containerName="extract-utilities" Dec 01 08:46:07 crc kubenswrapper[4873]: E1201 08:46:07.516759 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a18563f-c79c-4fed-83fb-5ee7d865e014" containerName="extract-content" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.516767 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a18563f-c79c-4fed-83fb-5ee7d865e014" containerName="extract-content" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.516917 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="73df40c6-5524-496f-92ba-36100c4af8bb" containerName="registry-server" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.516936 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a18563f-c79c-4fed-83fb-5ee7d865e014" containerName="registry-server" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.516944 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="d44d4877-74e9-4fdc-a062-92d031f1636d" containerName="registry-server" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.516960 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ea695f7-4e9e-4b12-8e0c-22431291576d" containerName="marketplace-operator" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.516967 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="bde4cec0-f82d-4724-9a01-f41a2bb3c93e" containerName="controller-manager" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.516978 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3fc6e9a-1447-41e2-8886-e5ea32e3d353" containerName="registry-server" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.516986 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ea695f7-4e9e-4b12-8e0c-22431291576d" containerName="marketplace-operator" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.517524 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7d57f65db7-hqchq" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.520656 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.525523 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.526615 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.527603 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.527660 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.531288 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.531836 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7d57f65db7-hqchq"] Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.532210 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.562235 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a735b178-621e-4cb2-a39e-562b631c3709-serving-cert\") pod \"controller-manager-7d57f65db7-hqchq\" (UID: \"a735b178-621e-4cb2-a39e-562b631c3709\") " pod="openshift-controller-manager/controller-manager-7d57f65db7-hqchq" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.562333 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a735b178-621e-4cb2-a39e-562b631c3709-client-ca\") pod \"controller-manager-7d57f65db7-hqchq\" (UID: \"a735b178-621e-4cb2-a39e-562b631c3709\") " pod="openshift-controller-manager/controller-manager-7d57f65db7-hqchq" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.562381 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a735b178-621e-4cb2-a39e-562b631c3709-config\") pod \"controller-manager-7d57f65db7-hqchq\" (UID: \"a735b178-621e-4cb2-a39e-562b631c3709\") " pod="openshift-controller-manager/controller-manager-7d57f65db7-hqchq" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.562465 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a735b178-621e-4cb2-a39e-562b631c3709-proxy-ca-bundles\") pod \"controller-manager-7d57f65db7-hqchq\" (UID: \"a735b178-621e-4cb2-a39e-562b631c3709\") " pod="openshift-controller-manager/controller-manager-7d57f65db7-hqchq" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.562504 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pmmt\" (UniqueName: \"kubernetes.io/projected/a735b178-621e-4cb2-a39e-562b631c3709-kube-api-access-7pmmt\") pod \"controller-manager-7d57f65db7-hqchq\" (UID: \"a735b178-621e-4cb2-a39e-562b631c3709\") " pod="openshift-controller-manager/controller-manager-7d57f65db7-hqchq" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.663931 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a735b178-621e-4cb2-a39e-562b631c3709-serving-cert\") pod \"controller-manager-7d57f65db7-hqchq\" (UID: \"a735b178-621e-4cb2-a39e-562b631c3709\") " pod="openshift-controller-manager/controller-manager-7d57f65db7-hqchq" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.664753 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a735b178-621e-4cb2-a39e-562b631c3709-client-ca\") pod \"controller-manager-7d57f65db7-hqchq\" (UID: \"a735b178-621e-4cb2-a39e-562b631c3709\") " pod="openshift-controller-manager/controller-manager-7d57f65db7-hqchq" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.664884 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a735b178-621e-4cb2-a39e-562b631c3709-config\") pod \"controller-manager-7d57f65db7-hqchq\" (UID: \"a735b178-621e-4cb2-a39e-562b631c3709\") " pod="openshift-controller-manager/controller-manager-7d57f65db7-hqchq" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.665047 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a735b178-621e-4cb2-a39e-562b631c3709-proxy-ca-bundles\") pod \"controller-manager-7d57f65db7-hqchq\" (UID: \"a735b178-621e-4cb2-a39e-562b631c3709\") " pod="openshift-controller-manager/controller-manager-7d57f65db7-hqchq" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.665187 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pmmt\" (UniqueName: \"kubernetes.io/projected/a735b178-621e-4cb2-a39e-562b631c3709-kube-api-access-7pmmt\") pod \"controller-manager-7d57f65db7-hqchq\" (UID: \"a735b178-621e-4cb2-a39e-562b631c3709\") " pod="openshift-controller-manager/controller-manager-7d57f65db7-hqchq" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.666078 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a735b178-621e-4cb2-a39e-562b631c3709-client-ca\") pod \"controller-manager-7d57f65db7-hqchq\" (UID: \"a735b178-621e-4cb2-a39e-562b631c3709\") " pod="openshift-controller-manager/controller-manager-7d57f65db7-hqchq" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.666473 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a735b178-621e-4cb2-a39e-562b631c3709-config\") pod \"controller-manager-7d57f65db7-hqchq\" (UID: \"a735b178-621e-4cb2-a39e-562b631c3709\") " pod="openshift-controller-manager/controller-manager-7d57f65db7-hqchq" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.666794 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a735b178-621e-4cb2-a39e-562b631c3709-proxy-ca-bundles\") pod \"controller-manager-7d57f65db7-hqchq\" (UID: \"a735b178-621e-4cb2-a39e-562b631c3709\") " pod="openshift-controller-manager/controller-manager-7d57f65db7-hqchq" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.675545 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a735b178-621e-4cb2-a39e-562b631c3709-serving-cert\") pod \"controller-manager-7d57f65db7-hqchq\" (UID: \"a735b178-621e-4cb2-a39e-562b631c3709\") " pod="openshift-controller-manager/controller-manager-7d57f65db7-hqchq" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.689822 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pmmt\" (UniqueName: \"kubernetes.io/projected/a735b178-621e-4cb2-a39e-562b631c3709-kube-api-access-7pmmt\") pod \"controller-manager-7d57f65db7-hqchq\" (UID: \"a735b178-621e-4cb2-a39e-562b631c3709\") " pod="openshift-controller-manager/controller-manager-7d57f65db7-hqchq" Dec 01 08:46:07 crc kubenswrapper[4873]: I1201 08:46:07.848034 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7d57f65db7-hqchq" Dec 01 08:46:08 crc kubenswrapper[4873]: I1201 08:46:08.060526 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7d57f65db7-hqchq"] Dec 01 08:46:08 crc kubenswrapper[4873]: I1201 08:46:08.437773 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bde4cec0-f82d-4724-9a01-f41a2bb3c93e" path="/var/lib/kubelet/pods/bde4cec0-f82d-4724-9a01-f41a2bb3c93e/volumes" Dec 01 08:46:08 crc kubenswrapper[4873]: I1201 08:46:08.992228 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7d57f65db7-hqchq" event={"ID":"a735b178-621e-4cb2-a39e-562b631c3709","Type":"ContainerStarted","Data":"f17f08bd65f02b66d311b73f64ba09b1a9661e6457d97fd694df6485afac092c"} Dec 01 08:46:08 crc kubenswrapper[4873]: I1201 08:46:08.992797 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7d57f65db7-hqchq" event={"ID":"a735b178-621e-4cb2-a39e-562b631c3709","Type":"ContainerStarted","Data":"1dd2f530adccb5be379e0fed424c026e5e642581413107519ba3d832b8c25506"} Dec 01 08:46:08 crc kubenswrapper[4873]: I1201 08:46:08.992825 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7d57f65db7-hqchq" Dec 01 08:46:08 crc kubenswrapper[4873]: I1201 08:46:08.997559 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7d57f65db7-hqchq" Dec 01 08:46:09 crc kubenswrapper[4873]: I1201 08:46:09.038226 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7d57f65db7-hqchq" podStartSLOduration=3.038196188 podStartE2EDuration="3.038196188s" podCreationTimestamp="2025-12-01 08:46:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:46:09.012121546 +0000 UTC m=+344.914230085" watchObservedRunningTime="2025-12-01 08:46:09.038196188 +0000 UTC m=+344.940304747" Dec 01 08:46:15 crc kubenswrapper[4873]: I1201 08:46:15.021999 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pdgt5"] Dec 01 08:46:15 crc kubenswrapper[4873]: I1201 08:46:15.024980 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pdgt5" Dec 01 08:46:15 crc kubenswrapper[4873]: I1201 08:46:15.027718 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 01 08:46:15 crc kubenswrapper[4873]: I1201 08:46:15.030310 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pdgt5"] Dec 01 08:46:15 crc kubenswrapper[4873]: I1201 08:46:15.068246 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96e1f659-b44b-4573-ba28-9605403487e5-utilities\") pod \"certified-operators-pdgt5\" (UID: \"96e1f659-b44b-4573-ba28-9605403487e5\") " pod="openshift-marketplace/certified-operators-pdgt5" Dec 01 08:46:15 crc kubenswrapper[4873]: I1201 08:46:15.068321 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96e1f659-b44b-4573-ba28-9605403487e5-catalog-content\") pod \"certified-operators-pdgt5\" (UID: \"96e1f659-b44b-4573-ba28-9605403487e5\") " pod="openshift-marketplace/certified-operators-pdgt5" Dec 01 08:46:15 crc kubenswrapper[4873]: I1201 08:46:15.068348 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbnhf\" (UniqueName: \"kubernetes.io/projected/96e1f659-b44b-4573-ba28-9605403487e5-kube-api-access-sbnhf\") pod \"certified-operators-pdgt5\" (UID: \"96e1f659-b44b-4573-ba28-9605403487e5\") " pod="openshift-marketplace/certified-operators-pdgt5" Dec 01 08:46:15 crc kubenswrapper[4873]: I1201 08:46:15.168884 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96e1f659-b44b-4573-ba28-9605403487e5-catalog-content\") pod \"certified-operators-pdgt5\" (UID: \"96e1f659-b44b-4573-ba28-9605403487e5\") " pod="openshift-marketplace/certified-operators-pdgt5" Dec 01 08:46:15 crc kubenswrapper[4873]: I1201 08:46:15.168947 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbnhf\" (UniqueName: \"kubernetes.io/projected/96e1f659-b44b-4573-ba28-9605403487e5-kube-api-access-sbnhf\") pod \"certified-operators-pdgt5\" (UID: \"96e1f659-b44b-4573-ba28-9605403487e5\") " pod="openshift-marketplace/certified-operators-pdgt5" Dec 01 08:46:15 crc kubenswrapper[4873]: I1201 08:46:15.169045 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96e1f659-b44b-4573-ba28-9605403487e5-utilities\") pod \"certified-operators-pdgt5\" (UID: \"96e1f659-b44b-4573-ba28-9605403487e5\") " pod="openshift-marketplace/certified-operators-pdgt5" Dec 01 08:46:15 crc kubenswrapper[4873]: I1201 08:46:15.169684 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96e1f659-b44b-4573-ba28-9605403487e5-catalog-content\") pod \"certified-operators-pdgt5\" (UID: \"96e1f659-b44b-4573-ba28-9605403487e5\") " pod="openshift-marketplace/certified-operators-pdgt5" Dec 01 08:46:15 crc kubenswrapper[4873]: I1201 08:46:15.169735 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96e1f659-b44b-4573-ba28-9605403487e5-utilities\") pod \"certified-operators-pdgt5\" (UID: \"96e1f659-b44b-4573-ba28-9605403487e5\") " pod="openshift-marketplace/certified-operators-pdgt5" Dec 01 08:46:15 crc kubenswrapper[4873]: I1201 08:46:15.193077 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6rsqz"] Dec 01 08:46:15 crc kubenswrapper[4873]: I1201 08:46:15.194311 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6rsqz" Dec 01 08:46:15 crc kubenswrapper[4873]: I1201 08:46:15.197814 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 01 08:46:15 crc kubenswrapper[4873]: I1201 08:46:15.198975 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbnhf\" (UniqueName: \"kubernetes.io/projected/96e1f659-b44b-4573-ba28-9605403487e5-kube-api-access-sbnhf\") pod \"certified-operators-pdgt5\" (UID: \"96e1f659-b44b-4573-ba28-9605403487e5\") " pod="openshift-marketplace/certified-operators-pdgt5" Dec 01 08:46:15 crc kubenswrapper[4873]: I1201 08:46:15.210908 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6rsqz"] Dec 01 08:46:15 crc kubenswrapper[4873]: I1201 08:46:15.270340 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c2e9631-829e-4ee3-af26-1f8bffc37029-utilities\") pod \"community-operators-6rsqz\" (UID: \"1c2e9631-829e-4ee3-af26-1f8bffc37029\") " pod="openshift-marketplace/community-operators-6rsqz" Dec 01 08:46:15 crc kubenswrapper[4873]: I1201 08:46:15.270495 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c2e9631-829e-4ee3-af26-1f8bffc37029-catalog-content\") pod \"community-operators-6rsqz\" (UID: \"1c2e9631-829e-4ee3-af26-1f8bffc37029\") " pod="openshift-marketplace/community-operators-6rsqz" Dec 01 08:46:15 crc kubenswrapper[4873]: I1201 08:46:15.270548 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8s284\" (UniqueName: \"kubernetes.io/projected/1c2e9631-829e-4ee3-af26-1f8bffc37029-kube-api-access-8s284\") pod \"community-operators-6rsqz\" (UID: \"1c2e9631-829e-4ee3-af26-1f8bffc37029\") " pod="openshift-marketplace/community-operators-6rsqz" Dec 01 08:46:15 crc kubenswrapper[4873]: I1201 08:46:15.355904 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pdgt5" Dec 01 08:46:15 crc kubenswrapper[4873]: I1201 08:46:15.371521 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c2e9631-829e-4ee3-af26-1f8bffc37029-utilities\") pod \"community-operators-6rsqz\" (UID: \"1c2e9631-829e-4ee3-af26-1f8bffc37029\") " pod="openshift-marketplace/community-operators-6rsqz" Dec 01 08:46:15 crc kubenswrapper[4873]: I1201 08:46:15.371599 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c2e9631-829e-4ee3-af26-1f8bffc37029-catalog-content\") pod \"community-operators-6rsqz\" (UID: \"1c2e9631-829e-4ee3-af26-1f8bffc37029\") " pod="openshift-marketplace/community-operators-6rsqz" Dec 01 08:46:15 crc kubenswrapper[4873]: I1201 08:46:15.371636 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8s284\" (UniqueName: \"kubernetes.io/projected/1c2e9631-829e-4ee3-af26-1f8bffc37029-kube-api-access-8s284\") pod \"community-operators-6rsqz\" (UID: \"1c2e9631-829e-4ee3-af26-1f8bffc37029\") " pod="openshift-marketplace/community-operators-6rsqz" Dec 01 08:46:15 crc kubenswrapper[4873]: I1201 08:46:15.372974 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c2e9631-829e-4ee3-af26-1f8bffc37029-utilities\") pod \"community-operators-6rsqz\" (UID: \"1c2e9631-829e-4ee3-af26-1f8bffc37029\") " pod="openshift-marketplace/community-operators-6rsqz" Dec 01 08:46:15 crc kubenswrapper[4873]: I1201 08:46:15.373245 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c2e9631-829e-4ee3-af26-1f8bffc37029-catalog-content\") pod \"community-operators-6rsqz\" (UID: \"1c2e9631-829e-4ee3-af26-1f8bffc37029\") " pod="openshift-marketplace/community-operators-6rsqz" Dec 01 08:46:15 crc kubenswrapper[4873]: I1201 08:46:15.401608 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8s284\" (UniqueName: \"kubernetes.io/projected/1c2e9631-829e-4ee3-af26-1f8bffc37029-kube-api-access-8s284\") pod \"community-operators-6rsqz\" (UID: \"1c2e9631-829e-4ee3-af26-1f8bffc37029\") " pod="openshift-marketplace/community-operators-6rsqz" Dec 01 08:46:15 crc kubenswrapper[4873]: I1201 08:46:15.544789 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6rsqz" Dec 01 08:46:15 crc kubenswrapper[4873]: I1201 08:46:15.808916 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pdgt5"] Dec 01 08:46:15 crc kubenswrapper[4873]: I1201 08:46:15.991325 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6rsqz"] Dec 01 08:46:16 crc kubenswrapper[4873]: I1201 08:46:16.035197 4873 generic.go:334] "Generic (PLEG): container finished" podID="96e1f659-b44b-4573-ba28-9605403487e5" containerID="b0b74481bb32a6d30908a11217df16c99e86703fe53ace58083d3e1d3a9e1259" exitCode=0 Dec 01 08:46:16 crc kubenswrapper[4873]: I1201 08:46:16.035244 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pdgt5" event={"ID":"96e1f659-b44b-4573-ba28-9605403487e5","Type":"ContainerDied","Data":"b0b74481bb32a6d30908a11217df16c99e86703fe53ace58083d3e1d3a9e1259"} Dec 01 08:46:16 crc kubenswrapper[4873]: I1201 08:46:16.035270 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pdgt5" event={"ID":"96e1f659-b44b-4573-ba28-9605403487e5","Type":"ContainerStarted","Data":"8b01227d1c7bde5e50b86ca6e4b533b29db90a59ec553c08460d3cf83f6f229e"} Dec 01 08:46:16 crc kubenswrapper[4873]: W1201 08:46:16.040042 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1c2e9631_829e_4ee3_af26_1f8bffc37029.slice/crio-fba2a7d144f7bb0c78e1ead6ce3bcc8858cad994e8aa7736d4b39e87fc5f6eda WatchSource:0}: Error finding container fba2a7d144f7bb0c78e1ead6ce3bcc8858cad994e8aa7736d4b39e87fc5f6eda: Status 404 returned error can't find the container with id fba2a7d144f7bb0c78e1ead6ce3bcc8858cad994e8aa7736d4b39e87fc5f6eda Dec 01 08:46:17 crc kubenswrapper[4873]: I1201 08:46:17.043630 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pdgt5" event={"ID":"96e1f659-b44b-4573-ba28-9605403487e5","Type":"ContainerStarted","Data":"1f04e9fc4f6be1a582ae0e47801c9c00cd2367ac410f5eb11b7bb526c5649cbf"} Dec 01 08:46:17 crc kubenswrapper[4873]: I1201 08:46:17.045551 4873 generic.go:334] "Generic (PLEG): container finished" podID="1c2e9631-829e-4ee3-af26-1f8bffc37029" containerID="6290fe887ebefaca070523485158812ab8b791bb81e722a43bc8de21c2f184e6" exitCode=0 Dec 01 08:46:17 crc kubenswrapper[4873]: I1201 08:46:17.045583 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6rsqz" event={"ID":"1c2e9631-829e-4ee3-af26-1f8bffc37029","Type":"ContainerDied","Data":"6290fe887ebefaca070523485158812ab8b791bb81e722a43bc8de21c2f184e6"} Dec 01 08:46:17 crc kubenswrapper[4873]: I1201 08:46:17.045601 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6rsqz" event={"ID":"1c2e9631-829e-4ee3-af26-1f8bffc37029","Type":"ContainerStarted","Data":"fba2a7d144f7bb0c78e1ead6ce3bcc8858cad994e8aa7736d4b39e87fc5f6eda"} Dec 01 08:46:17 crc kubenswrapper[4873]: I1201 08:46:17.398040 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-p98sp"] Dec 01 08:46:17 crc kubenswrapper[4873]: I1201 08:46:17.399438 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p98sp" Dec 01 08:46:17 crc kubenswrapper[4873]: I1201 08:46:17.403508 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 01 08:46:17 crc kubenswrapper[4873]: I1201 08:46:17.404424 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5czqb\" (UniqueName: \"kubernetes.io/projected/d62ca400-6749-4a3f-94ec-f7a5716584b9-kube-api-access-5czqb\") pod \"redhat-marketplace-p98sp\" (UID: \"d62ca400-6749-4a3f-94ec-f7a5716584b9\") " pod="openshift-marketplace/redhat-marketplace-p98sp" Dec 01 08:46:17 crc kubenswrapper[4873]: I1201 08:46:17.404526 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d62ca400-6749-4a3f-94ec-f7a5716584b9-catalog-content\") pod \"redhat-marketplace-p98sp\" (UID: \"d62ca400-6749-4a3f-94ec-f7a5716584b9\") " pod="openshift-marketplace/redhat-marketplace-p98sp" Dec 01 08:46:17 crc kubenswrapper[4873]: I1201 08:46:17.404572 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d62ca400-6749-4a3f-94ec-f7a5716584b9-utilities\") pod \"redhat-marketplace-p98sp\" (UID: \"d62ca400-6749-4a3f-94ec-f7a5716584b9\") " pod="openshift-marketplace/redhat-marketplace-p98sp" Dec 01 08:46:17 crc kubenswrapper[4873]: I1201 08:46:17.410350 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p98sp"] Dec 01 08:46:17 crc kubenswrapper[4873]: I1201 08:46:17.506082 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d62ca400-6749-4a3f-94ec-f7a5716584b9-catalog-content\") pod \"redhat-marketplace-p98sp\" (UID: \"d62ca400-6749-4a3f-94ec-f7a5716584b9\") " pod="openshift-marketplace/redhat-marketplace-p98sp" Dec 01 08:46:17 crc kubenswrapper[4873]: I1201 08:46:17.506176 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d62ca400-6749-4a3f-94ec-f7a5716584b9-utilities\") pod \"redhat-marketplace-p98sp\" (UID: \"d62ca400-6749-4a3f-94ec-f7a5716584b9\") " pod="openshift-marketplace/redhat-marketplace-p98sp" Dec 01 08:46:17 crc kubenswrapper[4873]: I1201 08:46:17.506212 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5czqb\" (UniqueName: \"kubernetes.io/projected/d62ca400-6749-4a3f-94ec-f7a5716584b9-kube-api-access-5czqb\") pod \"redhat-marketplace-p98sp\" (UID: \"d62ca400-6749-4a3f-94ec-f7a5716584b9\") " pod="openshift-marketplace/redhat-marketplace-p98sp" Dec 01 08:46:17 crc kubenswrapper[4873]: I1201 08:46:17.506887 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d62ca400-6749-4a3f-94ec-f7a5716584b9-catalog-content\") pod \"redhat-marketplace-p98sp\" (UID: \"d62ca400-6749-4a3f-94ec-f7a5716584b9\") " pod="openshift-marketplace/redhat-marketplace-p98sp" Dec 01 08:46:17 crc kubenswrapper[4873]: I1201 08:46:17.506921 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d62ca400-6749-4a3f-94ec-f7a5716584b9-utilities\") pod \"redhat-marketplace-p98sp\" (UID: \"d62ca400-6749-4a3f-94ec-f7a5716584b9\") " pod="openshift-marketplace/redhat-marketplace-p98sp" Dec 01 08:46:17 crc kubenswrapper[4873]: I1201 08:46:17.539487 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5czqb\" (UniqueName: \"kubernetes.io/projected/d62ca400-6749-4a3f-94ec-f7a5716584b9-kube-api-access-5czqb\") pod \"redhat-marketplace-p98sp\" (UID: \"d62ca400-6749-4a3f-94ec-f7a5716584b9\") " pod="openshift-marketplace/redhat-marketplace-p98sp" Dec 01 08:46:17 crc kubenswrapper[4873]: I1201 08:46:17.591235 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-v6fx2"] Dec 01 08:46:17 crc kubenswrapper[4873]: I1201 08:46:17.592357 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v6fx2" Dec 01 08:46:17 crc kubenswrapper[4873]: I1201 08:46:17.595539 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 01 08:46:17 crc kubenswrapper[4873]: I1201 08:46:17.603297 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-v6fx2"] Dec 01 08:46:17 crc kubenswrapper[4873]: I1201 08:46:17.608053 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e92483b-2415-4432-9059-3a60badb2e4a-utilities\") pod \"redhat-operators-v6fx2\" (UID: \"3e92483b-2415-4432-9059-3a60badb2e4a\") " pod="openshift-marketplace/redhat-operators-v6fx2" Dec 01 08:46:17 crc kubenswrapper[4873]: I1201 08:46:17.608127 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e92483b-2415-4432-9059-3a60badb2e4a-catalog-content\") pod \"redhat-operators-v6fx2\" (UID: \"3e92483b-2415-4432-9059-3a60badb2e4a\") " pod="openshift-marketplace/redhat-operators-v6fx2" Dec 01 08:46:17 crc kubenswrapper[4873]: I1201 08:46:17.608160 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chl86\" (UniqueName: \"kubernetes.io/projected/3e92483b-2415-4432-9059-3a60badb2e4a-kube-api-access-chl86\") pod \"redhat-operators-v6fx2\" (UID: \"3e92483b-2415-4432-9059-3a60badb2e4a\") " pod="openshift-marketplace/redhat-operators-v6fx2" Dec 01 08:46:17 crc kubenswrapper[4873]: I1201 08:46:17.709154 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e92483b-2415-4432-9059-3a60badb2e4a-catalog-content\") pod \"redhat-operators-v6fx2\" (UID: \"3e92483b-2415-4432-9059-3a60badb2e4a\") " pod="openshift-marketplace/redhat-operators-v6fx2" Dec 01 08:46:17 crc kubenswrapper[4873]: I1201 08:46:17.709236 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chl86\" (UniqueName: \"kubernetes.io/projected/3e92483b-2415-4432-9059-3a60badb2e4a-kube-api-access-chl86\") pod \"redhat-operators-v6fx2\" (UID: \"3e92483b-2415-4432-9059-3a60badb2e4a\") " pod="openshift-marketplace/redhat-operators-v6fx2" Dec 01 08:46:17 crc kubenswrapper[4873]: I1201 08:46:17.709291 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e92483b-2415-4432-9059-3a60badb2e4a-utilities\") pod \"redhat-operators-v6fx2\" (UID: \"3e92483b-2415-4432-9059-3a60badb2e4a\") " pod="openshift-marketplace/redhat-operators-v6fx2" Dec 01 08:46:17 crc kubenswrapper[4873]: I1201 08:46:17.710260 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e92483b-2415-4432-9059-3a60badb2e4a-catalog-content\") pod \"redhat-operators-v6fx2\" (UID: \"3e92483b-2415-4432-9059-3a60badb2e4a\") " pod="openshift-marketplace/redhat-operators-v6fx2" Dec 01 08:46:17 crc kubenswrapper[4873]: I1201 08:46:17.710309 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e92483b-2415-4432-9059-3a60badb2e4a-utilities\") pod \"redhat-operators-v6fx2\" (UID: \"3e92483b-2415-4432-9059-3a60badb2e4a\") " pod="openshift-marketplace/redhat-operators-v6fx2" Dec 01 08:46:17 crc kubenswrapper[4873]: I1201 08:46:17.713277 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p98sp" Dec 01 08:46:17 crc kubenswrapper[4873]: I1201 08:46:17.731494 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chl86\" (UniqueName: \"kubernetes.io/projected/3e92483b-2415-4432-9059-3a60badb2e4a-kube-api-access-chl86\") pod \"redhat-operators-v6fx2\" (UID: \"3e92483b-2415-4432-9059-3a60badb2e4a\") " pod="openshift-marketplace/redhat-operators-v6fx2" Dec 01 08:46:17 crc kubenswrapper[4873]: I1201 08:46:17.964910 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v6fx2" Dec 01 08:46:18 crc kubenswrapper[4873]: I1201 08:46:18.067618 4873 generic.go:334] "Generic (PLEG): container finished" podID="96e1f659-b44b-4573-ba28-9605403487e5" containerID="1f04e9fc4f6be1a582ae0e47801c9c00cd2367ac410f5eb11b7bb526c5649cbf" exitCode=0 Dec 01 08:46:18 crc kubenswrapper[4873]: I1201 08:46:18.067940 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pdgt5" event={"ID":"96e1f659-b44b-4573-ba28-9605403487e5","Type":"ContainerDied","Data":"1f04e9fc4f6be1a582ae0e47801c9c00cd2367ac410f5eb11b7bb526c5649cbf"} Dec 01 08:46:18 crc kubenswrapper[4873]: I1201 08:46:18.159758 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p98sp"] Dec 01 08:46:18 crc kubenswrapper[4873]: W1201 08:46:18.164171 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd62ca400_6749_4a3f_94ec_f7a5716584b9.slice/crio-5e5d4497e8b7da32d188945775d718737c3c592d59be520fc6999179d20cf7ff WatchSource:0}: Error finding container 5e5d4497e8b7da32d188945775d718737c3c592d59be520fc6999179d20cf7ff: Status 404 returned error can't find the container with id 5e5d4497e8b7da32d188945775d718737c3c592d59be520fc6999179d20cf7ff Dec 01 08:46:18 crc kubenswrapper[4873]: I1201 08:46:18.400956 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-v6fx2"] Dec 01 08:46:19 crc kubenswrapper[4873]: I1201 08:46:19.078927 4873 generic.go:334] "Generic (PLEG): container finished" podID="1c2e9631-829e-4ee3-af26-1f8bffc37029" containerID="0b93d4ae43167f894e68a1226cc5a2d30889d3b80794e27529f8ae3ac9284987" exitCode=0 Dec 01 08:46:19 crc kubenswrapper[4873]: I1201 08:46:19.079048 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6rsqz" event={"ID":"1c2e9631-829e-4ee3-af26-1f8bffc37029","Type":"ContainerDied","Data":"0b93d4ae43167f894e68a1226cc5a2d30889d3b80794e27529f8ae3ac9284987"} Dec 01 08:46:19 crc kubenswrapper[4873]: I1201 08:46:19.084806 4873 generic.go:334] "Generic (PLEG): container finished" podID="3e92483b-2415-4432-9059-3a60badb2e4a" containerID="d54f0d18b0d13c03e43a358fd40efc3635dcd0d46d6d2e3cf60e60bb63abbd0e" exitCode=0 Dec 01 08:46:19 crc kubenswrapper[4873]: I1201 08:46:19.084906 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v6fx2" event={"ID":"3e92483b-2415-4432-9059-3a60badb2e4a","Type":"ContainerDied","Data":"d54f0d18b0d13c03e43a358fd40efc3635dcd0d46d6d2e3cf60e60bb63abbd0e"} Dec 01 08:46:19 crc kubenswrapper[4873]: I1201 08:46:19.084958 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v6fx2" event={"ID":"3e92483b-2415-4432-9059-3a60badb2e4a","Type":"ContainerStarted","Data":"0f94694066fc8959f66ceba5b8aae3b723e53df5e64b5cf403aea5552565a283"} Dec 01 08:46:19 crc kubenswrapper[4873]: I1201 08:46:19.089693 4873 generic.go:334] "Generic (PLEG): container finished" podID="d62ca400-6749-4a3f-94ec-f7a5716584b9" containerID="f1e984446f9a6004d634af9c29f05d327bcd8ac01139b9ce73bced290f2473bc" exitCode=0 Dec 01 08:46:19 crc kubenswrapper[4873]: I1201 08:46:19.089890 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p98sp" event={"ID":"d62ca400-6749-4a3f-94ec-f7a5716584b9","Type":"ContainerDied","Data":"f1e984446f9a6004d634af9c29f05d327bcd8ac01139b9ce73bced290f2473bc"} Dec 01 08:46:19 crc kubenswrapper[4873]: I1201 08:46:19.090475 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p98sp" event={"ID":"d62ca400-6749-4a3f-94ec-f7a5716584b9","Type":"ContainerStarted","Data":"5e5d4497e8b7da32d188945775d718737c3c592d59be520fc6999179d20cf7ff"} Dec 01 08:46:19 crc kubenswrapper[4873]: I1201 08:46:19.099431 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pdgt5" event={"ID":"96e1f659-b44b-4573-ba28-9605403487e5","Type":"ContainerStarted","Data":"fe61383a8544782eff77b15312f9ead537aa713dfd5614a3a304d232ea242647"} Dec 01 08:46:19 crc kubenswrapper[4873]: I1201 08:46:19.126728 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pdgt5" podStartSLOduration=2.649962011 podStartE2EDuration="5.12670094s" podCreationTimestamp="2025-12-01 08:46:14 +0000 UTC" firstStartedPulling="2025-12-01 08:46:16.036657854 +0000 UTC m=+351.938766393" lastFinishedPulling="2025-12-01 08:46:18.513396783 +0000 UTC m=+354.415505322" observedRunningTime="2025-12-01 08:46:19.120403012 +0000 UTC m=+355.022511551" watchObservedRunningTime="2025-12-01 08:46:19.12670094 +0000 UTC m=+355.028809489" Dec 01 08:46:20 crc kubenswrapper[4873]: I1201 08:46:20.107544 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6rsqz" event={"ID":"1c2e9631-829e-4ee3-af26-1f8bffc37029","Type":"ContainerStarted","Data":"f165854cd4bc8b7a60aa520b4bdf503536cc30f8ac115ce50673b958ddd9c771"} Dec 01 08:46:20 crc kubenswrapper[4873]: I1201 08:46:20.112359 4873 generic.go:334] "Generic (PLEG): container finished" podID="d62ca400-6749-4a3f-94ec-f7a5716584b9" containerID="c8f38331591a3f32a6ceb0717c9ced43f74df46a0eb05d6016915d8186742acf" exitCode=0 Dec 01 08:46:20 crc kubenswrapper[4873]: I1201 08:46:20.112463 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p98sp" event={"ID":"d62ca400-6749-4a3f-94ec-f7a5716584b9","Type":"ContainerDied","Data":"c8f38331591a3f32a6ceb0717c9ced43f74df46a0eb05d6016915d8186742acf"} Dec 01 08:46:20 crc kubenswrapper[4873]: I1201 08:46:20.162124 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6rsqz" podStartSLOduration=2.581575482 podStartE2EDuration="5.161998658s" podCreationTimestamp="2025-12-01 08:46:15 +0000 UTC" firstStartedPulling="2025-12-01 08:46:17.047499168 +0000 UTC m=+352.949607707" lastFinishedPulling="2025-12-01 08:46:19.627922334 +0000 UTC m=+355.530030883" observedRunningTime="2025-12-01 08:46:20.136215303 +0000 UTC m=+356.038323842" watchObservedRunningTime="2025-12-01 08:46:20.161998658 +0000 UTC m=+356.064107197" Dec 01 08:46:21 crc kubenswrapper[4873]: I1201 08:46:21.121523 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p98sp" event={"ID":"d62ca400-6749-4a3f-94ec-f7a5716584b9","Type":"ContainerStarted","Data":"19541b3d8bab5f039a81898389a560fb310ec48120b5495911fbe43c66f28b08"} Dec 01 08:46:21 crc kubenswrapper[4873]: I1201 08:46:21.124498 4873 generic.go:334] "Generic (PLEG): container finished" podID="3e92483b-2415-4432-9059-3a60badb2e4a" containerID="282049e0f01a00f3a9f86b46e13ba8f0773b35e77e7bae3389fa5afb0345fe0f" exitCode=0 Dec 01 08:46:21 crc kubenswrapper[4873]: I1201 08:46:21.124622 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v6fx2" event={"ID":"3e92483b-2415-4432-9059-3a60badb2e4a","Type":"ContainerDied","Data":"282049e0f01a00f3a9f86b46e13ba8f0773b35e77e7bae3389fa5afb0345fe0f"} Dec 01 08:46:21 crc kubenswrapper[4873]: I1201 08:46:21.141882 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-p98sp" podStartSLOduration=2.533382283 podStartE2EDuration="4.141852586s" podCreationTimestamp="2025-12-01 08:46:17 +0000 UTC" firstStartedPulling="2025-12-01 08:46:19.09351922 +0000 UTC m=+354.995627749" lastFinishedPulling="2025-12-01 08:46:20.701989503 +0000 UTC m=+356.604098052" observedRunningTime="2025-12-01 08:46:21.140528595 +0000 UTC m=+357.042637134" watchObservedRunningTime="2025-12-01 08:46:21.141852586 +0000 UTC m=+357.043961165" Dec 01 08:46:22 crc kubenswrapper[4873]: I1201 08:46:22.131747 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v6fx2" event={"ID":"3e92483b-2415-4432-9059-3a60badb2e4a","Type":"ContainerStarted","Data":"77e9e5b01e9b977558eb33db73ab46412aba13efdd9c5f247fb6d7ade80ded1f"} Dec 01 08:46:22 crc kubenswrapper[4873]: I1201 08:46:22.155698 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-v6fx2" podStartSLOduration=2.708824724 podStartE2EDuration="5.155676381s" podCreationTimestamp="2025-12-01 08:46:17 +0000 UTC" firstStartedPulling="2025-12-01 08:46:19.089614519 +0000 UTC m=+354.991723068" lastFinishedPulling="2025-12-01 08:46:21.536466186 +0000 UTC m=+357.438574725" observedRunningTime="2025-12-01 08:46:22.151938373 +0000 UTC m=+358.054046912" watchObservedRunningTime="2025-12-01 08:46:22.155676381 +0000 UTC m=+358.057784920" Dec 01 08:46:25 crc kubenswrapper[4873]: I1201 08:46:25.357115 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-pdgt5" Dec 01 08:46:25 crc kubenswrapper[4873]: I1201 08:46:25.357199 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pdgt5" Dec 01 08:46:25 crc kubenswrapper[4873]: I1201 08:46:25.405236 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pdgt5" Dec 01 08:46:25 crc kubenswrapper[4873]: I1201 08:46:25.545558 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6rsqz" Dec 01 08:46:25 crc kubenswrapper[4873]: I1201 08:46:25.545635 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6rsqz" Dec 01 08:46:25 crc kubenswrapper[4873]: I1201 08:46:25.590833 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6rsqz" Dec 01 08:46:26 crc kubenswrapper[4873]: I1201 08:46:26.204405 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6rsqz" Dec 01 08:46:26 crc kubenswrapper[4873]: I1201 08:46:26.206519 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pdgt5" Dec 01 08:46:26 crc kubenswrapper[4873]: I1201 08:46:26.373994 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66649d5fd7-zcb7m"] Dec 01 08:46:26 crc kubenswrapper[4873]: I1201 08:46:26.374238 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-66649d5fd7-zcb7m" podUID="a6608aa5-ae02-444c-a9a7-b50ee7fa2bee" containerName="route-controller-manager" containerID="cri-o://faba54e86835a830f98929a8715e5f332c3e65db89554f46dde5af720cac5b97" gracePeriod=30 Dec 01 08:46:27 crc kubenswrapper[4873]: I1201 08:46:27.713871 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-p98sp" Dec 01 08:46:27 crc kubenswrapper[4873]: I1201 08:46:27.713949 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-p98sp" Dec 01 08:46:27 crc kubenswrapper[4873]: I1201 08:46:27.763376 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-p98sp" Dec 01 08:46:27 crc kubenswrapper[4873]: I1201 08:46:27.966468 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-v6fx2" Dec 01 08:46:27 crc kubenswrapper[4873]: I1201 08:46:27.966870 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-v6fx2" Dec 01 08:46:28 crc kubenswrapper[4873]: I1201 08:46:28.014228 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-v6fx2" Dec 01 08:46:28 crc kubenswrapper[4873]: I1201 08:46:28.173981 4873 generic.go:334] "Generic (PLEG): container finished" podID="a6608aa5-ae02-444c-a9a7-b50ee7fa2bee" containerID="faba54e86835a830f98929a8715e5f332c3e65db89554f46dde5af720cac5b97" exitCode=0 Dec 01 08:46:28 crc kubenswrapper[4873]: I1201 08:46:28.174067 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-66649d5fd7-zcb7m" event={"ID":"a6608aa5-ae02-444c-a9a7-b50ee7fa2bee","Type":"ContainerDied","Data":"faba54e86835a830f98929a8715e5f332c3e65db89554f46dde5af720cac5b97"} Dec 01 08:46:28 crc kubenswrapper[4873]: I1201 08:46:28.212054 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-p98sp" Dec 01 08:46:28 crc kubenswrapper[4873]: I1201 08:46:28.219275 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-v6fx2" Dec 01 08:46:28 crc kubenswrapper[4873]: I1201 08:46:28.694339 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-66649d5fd7-zcb7m" Dec 01 08:46:28 crc kubenswrapper[4873]: I1201 08:46:28.728337 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cddb77889-vmhsc"] Dec 01 08:46:28 crc kubenswrapper[4873]: E1201 08:46:28.728725 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6608aa5-ae02-444c-a9a7-b50ee7fa2bee" containerName="route-controller-manager" Dec 01 08:46:28 crc kubenswrapper[4873]: I1201 08:46:28.728743 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6608aa5-ae02-444c-a9a7-b50ee7fa2bee" containerName="route-controller-manager" Dec 01 08:46:28 crc kubenswrapper[4873]: I1201 08:46:28.728856 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6608aa5-ae02-444c-a9a7-b50ee7fa2bee" containerName="route-controller-manager" Dec 01 08:46:28 crc kubenswrapper[4873]: I1201 08:46:28.729358 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-cddb77889-vmhsc" Dec 01 08:46:28 crc kubenswrapper[4873]: I1201 08:46:28.746287 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cddb77889-vmhsc"] Dec 01 08:46:28 crc kubenswrapper[4873]: I1201 08:46:28.881881 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a6608aa5-ae02-444c-a9a7-b50ee7fa2bee-client-ca\") pod \"a6608aa5-ae02-444c-a9a7-b50ee7fa2bee\" (UID: \"a6608aa5-ae02-444c-a9a7-b50ee7fa2bee\") " Dec 01 08:46:28 crc kubenswrapper[4873]: I1201 08:46:28.882094 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6608aa5-ae02-444c-a9a7-b50ee7fa2bee-config\") pod \"a6608aa5-ae02-444c-a9a7-b50ee7fa2bee\" (UID: \"a6608aa5-ae02-444c-a9a7-b50ee7fa2bee\") " Dec 01 08:46:28 crc kubenswrapper[4873]: I1201 08:46:28.882230 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a6608aa5-ae02-444c-a9a7-b50ee7fa2bee-serving-cert\") pod \"a6608aa5-ae02-444c-a9a7-b50ee7fa2bee\" (UID: \"a6608aa5-ae02-444c-a9a7-b50ee7fa2bee\") " Dec 01 08:46:28 crc kubenswrapper[4873]: I1201 08:46:28.882302 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rdvqv\" (UniqueName: \"kubernetes.io/projected/a6608aa5-ae02-444c-a9a7-b50ee7fa2bee-kube-api-access-rdvqv\") pod \"a6608aa5-ae02-444c-a9a7-b50ee7fa2bee\" (UID: \"a6608aa5-ae02-444c-a9a7-b50ee7fa2bee\") " Dec 01 08:46:28 crc kubenswrapper[4873]: I1201 08:46:28.882541 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/08d69e9e-d868-4377-8a4d-e29306841cf2-serving-cert\") pod \"route-controller-manager-cddb77889-vmhsc\" (UID: \"08d69e9e-d868-4377-8a4d-e29306841cf2\") " pod="openshift-route-controller-manager/route-controller-manager-cddb77889-vmhsc" Dec 01 08:46:28 crc kubenswrapper[4873]: I1201 08:46:28.882642 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/08d69e9e-d868-4377-8a4d-e29306841cf2-client-ca\") pod \"route-controller-manager-cddb77889-vmhsc\" (UID: \"08d69e9e-d868-4377-8a4d-e29306841cf2\") " pod="openshift-route-controller-manager/route-controller-manager-cddb77889-vmhsc" Dec 01 08:46:28 crc kubenswrapper[4873]: I1201 08:46:28.882725 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08d69e9e-d868-4377-8a4d-e29306841cf2-config\") pod \"route-controller-manager-cddb77889-vmhsc\" (UID: \"08d69e9e-d868-4377-8a4d-e29306841cf2\") " pod="openshift-route-controller-manager/route-controller-manager-cddb77889-vmhsc" Dec 01 08:46:28 crc kubenswrapper[4873]: I1201 08:46:28.882758 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x525h\" (UniqueName: \"kubernetes.io/projected/08d69e9e-d868-4377-8a4d-e29306841cf2-kube-api-access-x525h\") pod \"route-controller-manager-cddb77889-vmhsc\" (UID: \"08d69e9e-d868-4377-8a4d-e29306841cf2\") " pod="openshift-route-controller-manager/route-controller-manager-cddb77889-vmhsc" Dec 01 08:46:28 crc kubenswrapper[4873]: I1201 08:46:28.883198 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6608aa5-ae02-444c-a9a7-b50ee7fa2bee-client-ca" (OuterVolumeSpecName: "client-ca") pod "a6608aa5-ae02-444c-a9a7-b50ee7fa2bee" (UID: "a6608aa5-ae02-444c-a9a7-b50ee7fa2bee"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:46:28 crc kubenswrapper[4873]: I1201 08:46:28.883319 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6608aa5-ae02-444c-a9a7-b50ee7fa2bee-config" (OuterVolumeSpecName: "config") pod "a6608aa5-ae02-444c-a9a7-b50ee7fa2bee" (UID: "a6608aa5-ae02-444c-a9a7-b50ee7fa2bee"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:46:28 crc kubenswrapper[4873]: I1201 08:46:28.888760 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6608aa5-ae02-444c-a9a7-b50ee7fa2bee-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a6608aa5-ae02-444c-a9a7-b50ee7fa2bee" (UID: "a6608aa5-ae02-444c-a9a7-b50ee7fa2bee"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:46:28 crc kubenswrapper[4873]: I1201 08:46:28.888978 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6608aa5-ae02-444c-a9a7-b50ee7fa2bee-kube-api-access-rdvqv" (OuterVolumeSpecName: "kube-api-access-rdvqv") pod "a6608aa5-ae02-444c-a9a7-b50ee7fa2bee" (UID: "a6608aa5-ae02-444c-a9a7-b50ee7fa2bee"). InnerVolumeSpecName "kube-api-access-rdvqv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:46:28 crc kubenswrapper[4873]: I1201 08:46:28.984681 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08d69e9e-d868-4377-8a4d-e29306841cf2-config\") pod \"route-controller-manager-cddb77889-vmhsc\" (UID: \"08d69e9e-d868-4377-8a4d-e29306841cf2\") " pod="openshift-route-controller-manager/route-controller-manager-cddb77889-vmhsc" Dec 01 08:46:28 crc kubenswrapper[4873]: I1201 08:46:28.984745 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x525h\" (UniqueName: \"kubernetes.io/projected/08d69e9e-d868-4377-8a4d-e29306841cf2-kube-api-access-x525h\") pod \"route-controller-manager-cddb77889-vmhsc\" (UID: \"08d69e9e-d868-4377-8a4d-e29306841cf2\") " pod="openshift-route-controller-manager/route-controller-manager-cddb77889-vmhsc" Dec 01 08:46:28 crc kubenswrapper[4873]: I1201 08:46:28.984784 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/08d69e9e-d868-4377-8a4d-e29306841cf2-serving-cert\") pod \"route-controller-manager-cddb77889-vmhsc\" (UID: \"08d69e9e-d868-4377-8a4d-e29306841cf2\") " pod="openshift-route-controller-manager/route-controller-manager-cddb77889-vmhsc" Dec 01 08:46:28 crc kubenswrapper[4873]: I1201 08:46:28.984890 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/08d69e9e-d868-4377-8a4d-e29306841cf2-client-ca\") pod \"route-controller-manager-cddb77889-vmhsc\" (UID: \"08d69e9e-d868-4377-8a4d-e29306841cf2\") " pod="openshift-route-controller-manager/route-controller-manager-cddb77889-vmhsc" Dec 01 08:46:28 crc kubenswrapper[4873]: I1201 08:46:28.984985 4873 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6608aa5-ae02-444c-a9a7-b50ee7fa2bee-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:46:28 crc kubenswrapper[4873]: I1201 08:46:28.985031 4873 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a6608aa5-ae02-444c-a9a7-b50ee7fa2bee-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:46:28 crc kubenswrapper[4873]: I1201 08:46:28.985046 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rdvqv\" (UniqueName: \"kubernetes.io/projected/a6608aa5-ae02-444c-a9a7-b50ee7fa2bee-kube-api-access-rdvqv\") on node \"crc\" DevicePath \"\"" Dec 01 08:46:28 crc kubenswrapper[4873]: I1201 08:46:28.985061 4873 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a6608aa5-ae02-444c-a9a7-b50ee7fa2bee-client-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:46:28 crc kubenswrapper[4873]: I1201 08:46:28.986189 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/08d69e9e-d868-4377-8a4d-e29306841cf2-client-ca\") pod \"route-controller-manager-cddb77889-vmhsc\" (UID: \"08d69e9e-d868-4377-8a4d-e29306841cf2\") " pod="openshift-route-controller-manager/route-controller-manager-cddb77889-vmhsc" Dec 01 08:46:28 crc kubenswrapper[4873]: I1201 08:46:28.986225 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08d69e9e-d868-4377-8a4d-e29306841cf2-config\") pod \"route-controller-manager-cddb77889-vmhsc\" (UID: \"08d69e9e-d868-4377-8a4d-e29306841cf2\") " pod="openshift-route-controller-manager/route-controller-manager-cddb77889-vmhsc" Dec 01 08:46:28 crc kubenswrapper[4873]: I1201 08:46:28.992941 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/08d69e9e-d868-4377-8a4d-e29306841cf2-serving-cert\") pod \"route-controller-manager-cddb77889-vmhsc\" (UID: \"08d69e9e-d868-4377-8a4d-e29306841cf2\") " pod="openshift-route-controller-manager/route-controller-manager-cddb77889-vmhsc" Dec 01 08:46:29 crc kubenswrapper[4873]: I1201 08:46:29.003829 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x525h\" (UniqueName: \"kubernetes.io/projected/08d69e9e-d868-4377-8a4d-e29306841cf2-kube-api-access-x525h\") pod \"route-controller-manager-cddb77889-vmhsc\" (UID: \"08d69e9e-d868-4377-8a4d-e29306841cf2\") " pod="openshift-route-controller-manager/route-controller-manager-cddb77889-vmhsc" Dec 01 08:46:29 crc kubenswrapper[4873]: I1201 08:46:29.054228 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-cddb77889-vmhsc" Dec 01 08:46:29 crc kubenswrapper[4873]: I1201 08:46:29.185247 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-66649d5fd7-zcb7m" event={"ID":"a6608aa5-ae02-444c-a9a7-b50ee7fa2bee","Type":"ContainerDied","Data":"763e98f8a9bb25c1c72a66d258aa1fcc43dc66b48dcd2adde7335927a3904b65"} Dec 01 08:46:29 crc kubenswrapper[4873]: I1201 08:46:29.185319 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-66649d5fd7-zcb7m" Dec 01 08:46:29 crc kubenswrapper[4873]: I1201 08:46:29.185330 4873 scope.go:117] "RemoveContainer" containerID="faba54e86835a830f98929a8715e5f332c3e65db89554f46dde5af720cac5b97" Dec 01 08:46:29 crc kubenswrapper[4873]: I1201 08:46:29.226745 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66649d5fd7-zcb7m"] Dec 01 08:46:29 crc kubenswrapper[4873]: I1201 08:46:29.230718 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66649d5fd7-zcb7m"] Dec 01 08:46:29 crc kubenswrapper[4873]: I1201 08:46:29.549559 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cddb77889-vmhsc"] Dec 01 08:46:30 crc kubenswrapper[4873]: I1201 08:46:30.194398 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-cddb77889-vmhsc" event={"ID":"08d69e9e-d868-4377-8a4d-e29306841cf2","Type":"ContainerStarted","Data":"4baa4ecc527cca0b103f1f7dc84c3f34b4babf24e0d8748a6c6f141f97f88470"} Dec 01 08:46:30 crc kubenswrapper[4873]: I1201 08:46:30.194911 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-cddb77889-vmhsc" event={"ID":"08d69e9e-d868-4377-8a4d-e29306841cf2","Type":"ContainerStarted","Data":"c3b77b61cd4e5620c88b44fa93a55e3d805e262decc66e67470c2f75d51577a6"} Dec 01 08:46:30 crc kubenswrapper[4873]: I1201 08:46:30.194938 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-cddb77889-vmhsc" Dec 01 08:46:30 crc kubenswrapper[4873]: I1201 08:46:30.200437 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-cddb77889-vmhsc" Dec 01 08:46:30 crc kubenswrapper[4873]: I1201 08:46:30.216218 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-cddb77889-vmhsc" podStartSLOduration=4.216191725 podStartE2EDuration="4.216191725s" podCreationTimestamp="2025-12-01 08:46:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:46:30.213986863 +0000 UTC m=+366.116095412" watchObservedRunningTime="2025-12-01 08:46:30.216191725 +0000 UTC m=+366.118300264" Dec 01 08:46:30 crc kubenswrapper[4873]: I1201 08:46:30.437269 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6608aa5-ae02-444c-a9a7-b50ee7fa2bee" path="/var/lib/kubelet/pods/a6608aa5-ae02-444c-a9a7-b50ee7fa2bee/volumes" Dec 01 08:46:31 crc kubenswrapper[4873]: I1201 08:46:31.059174 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:46:31 crc kubenswrapper[4873]: I1201 08:46:31.059285 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:46:35 crc kubenswrapper[4873]: I1201 08:46:35.409923 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-dxl9b"] Dec 01 08:46:35 crc kubenswrapper[4873]: I1201 08:46:35.411681 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-dxl9b" Dec 01 08:46:35 crc kubenswrapper[4873]: I1201 08:46:35.440610 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-dxl9b"] Dec 01 08:46:35 crc kubenswrapper[4873]: I1201 08:46:35.596596 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1b11a04b-782f-4c20-a7df-e2e172921002-ca-trust-extracted\") pod \"image-registry-66df7c8f76-dxl9b\" (UID: \"1b11a04b-782f-4c20-a7df-e2e172921002\") " pod="openshift-image-registry/image-registry-66df7c8f76-dxl9b" Dec 01 08:46:35 crc kubenswrapper[4873]: I1201 08:46:35.596673 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1b11a04b-782f-4c20-a7df-e2e172921002-bound-sa-token\") pod \"image-registry-66df7c8f76-dxl9b\" (UID: \"1b11a04b-782f-4c20-a7df-e2e172921002\") " pod="openshift-image-registry/image-registry-66df7c8f76-dxl9b" Dec 01 08:46:35 crc kubenswrapper[4873]: I1201 08:46:35.596697 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsq7j\" (UniqueName: \"kubernetes.io/projected/1b11a04b-782f-4c20-a7df-e2e172921002-kube-api-access-rsq7j\") pod \"image-registry-66df7c8f76-dxl9b\" (UID: \"1b11a04b-782f-4c20-a7df-e2e172921002\") " pod="openshift-image-registry/image-registry-66df7c8f76-dxl9b" Dec 01 08:46:35 crc kubenswrapper[4873]: I1201 08:46:35.596788 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-dxl9b\" (UID: \"1b11a04b-782f-4c20-a7df-e2e172921002\") " pod="openshift-image-registry/image-registry-66df7c8f76-dxl9b" Dec 01 08:46:35 crc kubenswrapper[4873]: I1201 08:46:35.596920 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1b11a04b-782f-4c20-a7df-e2e172921002-installation-pull-secrets\") pod \"image-registry-66df7c8f76-dxl9b\" (UID: \"1b11a04b-782f-4c20-a7df-e2e172921002\") " pod="openshift-image-registry/image-registry-66df7c8f76-dxl9b" Dec 01 08:46:35 crc kubenswrapper[4873]: I1201 08:46:35.597160 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1b11a04b-782f-4c20-a7df-e2e172921002-registry-tls\") pod \"image-registry-66df7c8f76-dxl9b\" (UID: \"1b11a04b-782f-4c20-a7df-e2e172921002\") " pod="openshift-image-registry/image-registry-66df7c8f76-dxl9b" Dec 01 08:46:35 crc kubenswrapper[4873]: I1201 08:46:35.597317 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1b11a04b-782f-4c20-a7df-e2e172921002-trusted-ca\") pod \"image-registry-66df7c8f76-dxl9b\" (UID: \"1b11a04b-782f-4c20-a7df-e2e172921002\") " pod="openshift-image-registry/image-registry-66df7c8f76-dxl9b" Dec 01 08:46:35 crc kubenswrapper[4873]: I1201 08:46:35.597542 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1b11a04b-782f-4c20-a7df-e2e172921002-registry-certificates\") pod \"image-registry-66df7c8f76-dxl9b\" (UID: \"1b11a04b-782f-4c20-a7df-e2e172921002\") " pod="openshift-image-registry/image-registry-66df7c8f76-dxl9b" Dec 01 08:46:35 crc kubenswrapper[4873]: I1201 08:46:35.628185 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-dxl9b\" (UID: \"1b11a04b-782f-4c20-a7df-e2e172921002\") " pod="openshift-image-registry/image-registry-66df7c8f76-dxl9b" Dec 01 08:46:35 crc kubenswrapper[4873]: I1201 08:46:35.699499 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1b11a04b-782f-4c20-a7df-e2e172921002-registry-certificates\") pod \"image-registry-66df7c8f76-dxl9b\" (UID: \"1b11a04b-782f-4c20-a7df-e2e172921002\") " pod="openshift-image-registry/image-registry-66df7c8f76-dxl9b" Dec 01 08:46:35 crc kubenswrapper[4873]: I1201 08:46:35.699558 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1b11a04b-782f-4c20-a7df-e2e172921002-ca-trust-extracted\") pod \"image-registry-66df7c8f76-dxl9b\" (UID: \"1b11a04b-782f-4c20-a7df-e2e172921002\") " pod="openshift-image-registry/image-registry-66df7c8f76-dxl9b" Dec 01 08:46:35 crc kubenswrapper[4873]: I1201 08:46:35.699592 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1b11a04b-782f-4c20-a7df-e2e172921002-bound-sa-token\") pod \"image-registry-66df7c8f76-dxl9b\" (UID: \"1b11a04b-782f-4c20-a7df-e2e172921002\") " pod="openshift-image-registry/image-registry-66df7c8f76-dxl9b" Dec 01 08:46:35 crc kubenswrapper[4873]: I1201 08:46:35.699621 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsq7j\" (UniqueName: \"kubernetes.io/projected/1b11a04b-782f-4c20-a7df-e2e172921002-kube-api-access-rsq7j\") pod \"image-registry-66df7c8f76-dxl9b\" (UID: \"1b11a04b-782f-4c20-a7df-e2e172921002\") " pod="openshift-image-registry/image-registry-66df7c8f76-dxl9b" Dec 01 08:46:35 crc kubenswrapper[4873]: I1201 08:46:35.699659 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1b11a04b-782f-4c20-a7df-e2e172921002-installation-pull-secrets\") pod \"image-registry-66df7c8f76-dxl9b\" (UID: \"1b11a04b-782f-4c20-a7df-e2e172921002\") " pod="openshift-image-registry/image-registry-66df7c8f76-dxl9b" Dec 01 08:46:35 crc kubenswrapper[4873]: I1201 08:46:35.699693 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1b11a04b-782f-4c20-a7df-e2e172921002-registry-tls\") pod \"image-registry-66df7c8f76-dxl9b\" (UID: \"1b11a04b-782f-4c20-a7df-e2e172921002\") " pod="openshift-image-registry/image-registry-66df7c8f76-dxl9b" Dec 01 08:46:35 crc kubenswrapper[4873]: I1201 08:46:35.699717 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1b11a04b-782f-4c20-a7df-e2e172921002-trusted-ca\") pod \"image-registry-66df7c8f76-dxl9b\" (UID: \"1b11a04b-782f-4c20-a7df-e2e172921002\") " pod="openshift-image-registry/image-registry-66df7c8f76-dxl9b" Dec 01 08:46:35 crc kubenswrapper[4873]: I1201 08:46:35.700541 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1b11a04b-782f-4c20-a7df-e2e172921002-ca-trust-extracted\") pod \"image-registry-66df7c8f76-dxl9b\" (UID: \"1b11a04b-782f-4c20-a7df-e2e172921002\") " pod="openshift-image-registry/image-registry-66df7c8f76-dxl9b" Dec 01 08:46:35 crc kubenswrapper[4873]: I1201 08:46:35.701382 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1b11a04b-782f-4c20-a7df-e2e172921002-trusted-ca\") pod \"image-registry-66df7c8f76-dxl9b\" (UID: \"1b11a04b-782f-4c20-a7df-e2e172921002\") " pod="openshift-image-registry/image-registry-66df7c8f76-dxl9b" Dec 01 08:46:35 crc kubenswrapper[4873]: I1201 08:46:35.702836 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1b11a04b-782f-4c20-a7df-e2e172921002-registry-certificates\") pod \"image-registry-66df7c8f76-dxl9b\" (UID: \"1b11a04b-782f-4c20-a7df-e2e172921002\") " pod="openshift-image-registry/image-registry-66df7c8f76-dxl9b" Dec 01 08:46:35 crc kubenswrapper[4873]: I1201 08:46:35.709746 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1b11a04b-782f-4c20-a7df-e2e172921002-installation-pull-secrets\") pod \"image-registry-66df7c8f76-dxl9b\" (UID: \"1b11a04b-782f-4c20-a7df-e2e172921002\") " pod="openshift-image-registry/image-registry-66df7c8f76-dxl9b" Dec 01 08:46:35 crc kubenswrapper[4873]: I1201 08:46:35.709806 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1b11a04b-782f-4c20-a7df-e2e172921002-registry-tls\") pod \"image-registry-66df7c8f76-dxl9b\" (UID: \"1b11a04b-782f-4c20-a7df-e2e172921002\") " pod="openshift-image-registry/image-registry-66df7c8f76-dxl9b" Dec 01 08:46:35 crc kubenswrapper[4873]: I1201 08:46:35.721893 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1b11a04b-782f-4c20-a7df-e2e172921002-bound-sa-token\") pod \"image-registry-66df7c8f76-dxl9b\" (UID: \"1b11a04b-782f-4c20-a7df-e2e172921002\") " pod="openshift-image-registry/image-registry-66df7c8f76-dxl9b" Dec 01 08:46:35 crc kubenswrapper[4873]: I1201 08:46:35.726041 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsq7j\" (UniqueName: \"kubernetes.io/projected/1b11a04b-782f-4c20-a7df-e2e172921002-kube-api-access-rsq7j\") pod \"image-registry-66df7c8f76-dxl9b\" (UID: \"1b11a04b-782f-4c20-a7df-e2e172921002\") " pod="openshift-image-registry/image-registry-66df7c8f76-dxl9b" Dec 01 08:46:35 crc kubenswrapper[4873]: I1201 08:46:35.759337 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-dxl9b" Dec 01 08:46:36 crc kubenswrapper[4873]: I1201 08:46:36.292662 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-dxl9b"] Dec 01 08:46:36 crc kubenswrapper[4873]: W1201 08:46:36.297454 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1b11a04b_782f_4c20_a7df_e2e172921002.slice/crio-164319b1a5b86ac77b4feecec32da17d841c70b7e3537eb8c02e997fd0ab1f10 WatchSource:0}: Error finding container 164319b1a5b86ac77b4feecec32da17d841c70b7e3537eb8c02e997fd0ab1f10: Status 404 returned error can't find the container with id 164319b1a5b86ac77b4feecec32da17d841c70b7e3537eb8c02e997fd0ab1f10 Dec 01 08:46:37 crc kubenswrapper[4873]: I1201 08:46:37.241445 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-dxl9b" event={"ID":"1b11a04b-782f-4c20-a7df-e2e172921002","Type":"ContainerStarted","Data":"8f5c448030241b966d605669f46cc96855f50b4d6d43d6f955c35e87d69a3bbd"} Dec 01 08:46:37 crc kubenswrapper[4873]: I1201 08:46:37.241894 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-dxl9b" event={"ID":"1b11a04b-782f-4c20-a7df-e2e172921002","Type":"ContainerStarted","Data":"164319b1a5b86ac77b4feecec32da17d841c70b7e3537eb8c02e997fd0ab1f10"} Dec 01 08:46:37 crc kubenswrapper[4873]: I1201 08:46:37.241918 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-dxl9b" Dec 01 08:46:37 crc kubenswrapper[4873]: I1201 08:46:37.267883 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-dxl9b" podStartSLOduration=2.267850651 podStartE2EDuration="2.267850651s" podCreationTimestamp="2025-12-01 08:46:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:46:37.262802152 +0000 UTC m=+373.164910691" watchObservedRunningTime="2025-12-01 08:46:37.267850651 +0000 UTC m=+373.169959190" Dec 01 08:46:55 crc kubenswrapper[4873]: I1201 08:46:55.770182 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-dxl9b" Dec 01 08:46:55 crc kubenswrapper[4873]: I1201 08:46:55.855813 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cjn5c"] Dec 01 08:47:01 crc kubenswrapper[4873]: I1201 08:47:01.059751 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:47:01 crc kubenswrapper[4873]: I1201 08:47:01.060386 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:47:01 crc kubenswrapper[4873]: I1201 08:47:01.060457 4873 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" Dec 01 08:47:01 crc kubenswrapper[4873]: I1201 08:47:01.061115 4873 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9c30061eb3de4736a094430a40ceb3b59f45bd35deebb43c4c6aac312bafc4a8"} pod="openshift-machine-config-operator/machine-config-daemon-scwpp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 08:47:01 crc kubenswrapper[4873]: I1201 08:47:01.061190 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" containerID="cri-o://9c30061eb3de4736a094430a40ceb3b59f45bd35deebb43c4c6aac312bafc4a8" gracePeriod=600 Dec 01 08:47:01 crc kubenswrapper[4873]: I1201 08:47:01.613531 4873 generic.go:334] "Generic (PLEG): container finished" podID="fef7b114-0e07-402d-a37b-315c36011f4b" containerID="9c30061eb3de4736a094430a40ceb3b59f45bd35deebb43c4c6aac312bafc4a8" exitCode=0 Dec 01 08:47:01 crc kubenswrapper[4873]: I1201 08:47:01.613625 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" event={"ID":"fef7b114-0e07-402d-a37b-315c36011f4b","Type":"ContainerDied","Data":"9c30061eb3de4736a094430a40ceb3b59f45bd35deebb43c4c6aac312bafc4a8"} Dec 01 08:47:01 crc kubenswrapper[4873]: I1201 08:47:01.614203 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" event={"ID":"fef7b114-0e07-402d-a37b-315c36011f4b","Type":"ContainerStarted","Data":"0d13b8cddab09d2b96e4bfebd3f2cc00fd3cfd06d73474412b9d0387a8e1d7e0"} Dec 01 08:47:01 crc kubenswrapper[4873]: I1201 08:47:01.614276 4873 scope.go:117] "RemoveContainer" containerID="b67d090dbd4e672a6d3e777da244480eaf270518c68f8eaf2292ef7c366b8da4" Dec 01 08:47:20 crc kubenswrapper[4873]: I1201 08:47:20.904771 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" podUID="952ae5ba-156f-41a7-950a-1783e2e69f94" containerName="registry" containerID="cri-o://be35e87e40c6a12a9a8bf6f80153adc4d38b8d8466027397bc51f10e7db3f1c0" gracePeriod=30 Dec 01 08:47:21 crc kubenswrapper[4873]: I1201 08:47:21.431973 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:47:21 crc kubenswrapper[4873]: I1201 08:47:21.595120 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/952ae5ba-156f-41a7-950a-1783e2e69f94-trusted-ca\") pod \"952ae5ba-156f-41a7-950a-1783e2e69f94\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " Dec 01 08:47:21 crc kubenswrapper[4873]: I1201 08:47:21.595216 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/952ae5ba-156f-41a7-950a-1783e2e69f94-ca-trust-extracted\") pod \"952ae5ba-156f-41a7-950a-1783e2e69f94\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " Dec 01 08:47:21 crc kubenswrapper[4873]: I1201 08:47:21.595261 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/952ae5ba-156f-41a7-950a-1783e2e69f94-registry-tls\") pod \"952ae5ba-156f-41a7-950a-1783e2e69f94\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " Dec 01 08:47:21 crc kubenswrapper[4873]: I1201 08:47:21.595321 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qlc6g\" (UniqueName: \"kubernetes.io/projected/952ae5ba-156f-41a7-950a-1783e2e69f94-kube-api-access-qlc6g\") pod \"952ae5ba-156f-41a7-950a-1783e2e69f94\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " Dec 01 08:47:21 crc kubenswrapper[4873]: I1201 08:47:21.595363 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/952ae5ba-156f-41a7-950a-1783e2e69f94-bound-sa-token\") pod \"952ae5ba-156f-41a7-950a-1783e2e69f94\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " Dec 01 08:47:21 crc kubenswrapper[4873]: I1201 08:47:21.595410 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/952ae5ba-156f-41a7-950a-1783e2e69f94-registry-certificates\") pod \"952ae5ba-156f-41a7-950a-1783e2e69f94\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " Dec 01 08:47:21 crc kubenswrapper[4873]: I1201 08:47:21.595441 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/952ae5ba-156f-41a7-950a-1783e2e69f94-installation-pull-secrets\") pod \"952ae5ba-156f-41a7-950a-1783e2e69f94\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " Dec 01 08:47:21 crc kubenswrapper[4873]: I1201 08:47:21.595630 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"952ae5ba-156f-41a7-950a-1783e2e69f94\" (UID: \"952ae5ba-156f-41a7-950a-1783e2e69f94\") " Dec 01 08:47:21 crc kubenswrapper[4873]: I1201 08:47:21.596796 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/952ae5ba-156f-41a7-950a-1783e2e69f94-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "952ae5ba-156f-41a7-950a-1783e2e69f94" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:47:21 crc kubenswrapper[4873]: I1201 08:47:21.596842 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/952ae5ba-156f-41a7-950a-1783e2e69f94-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "952ae5ba-156f-41a7-950a-1783e2e69f94" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:47:21 crc kubenswrapper[4873]: I1201 08:47:21.603584 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/952ae5ba-156f-41a7-950a-1783e2e69f94-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "952ae5ba-156f-41a7-950a-1783e2e69f94" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:47:21 crc kubenswrapper[4873]: I1201 08:47:21.604316 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/952ae5ba-156f-41a7-950a-1783e2e69f94-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "952ae5ba-156f-41a7-950a-1783e2e69f94" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:47:21 crc kubenswrapper[4873]: I1201 08:47:21.604810 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/952ae5ba-156f-41a7-950a-1783e2e69f94-kube-api-access-qlc6g" (OuterVolumeSpecName: "kube-api-access-qlc6g") pod "952ae5ba-156f-41a7-950a-1783e2e69f94" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94"). InnerVolumeSpecName "kube-api-access-qlc6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:47:21 crc kubenswrapper[4873]: I1201 08:47:21.605199 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/952ae5ba-156f-41a7-950a-1783e2e69f94-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "952ae5ba-156f-41a7-950a-1783e2e69f94" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:47:21 crc kubenswrapper[4873]: I1201 08:47:21.608762 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "952ae5ba-156f-41a7-950a-1783e2e69f94" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 01 08:47:21 crc kubenswrapper[4873]: I1201 08:47:21.615152 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/952ae5ba-156f-41a7-950a-1783e2e69f94-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "952ae5ba-156f-41a7-950a-1783e2e69f94" (UID: "952ae5ba-156f-41a7-950a-1783e2e69f94"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:47:21 crc kubenswrapper[4873]: I1201 08:47:21.697234 4873 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/952ae5ba-156f-41a7-950a-1783e2e69f94-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 01 08:47:21 crc kubenswrapper[4873]: I1201 08:47:21.697273 4873 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/952ae5ba-156f-41a7-950a-1783e2e69f94-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 01 08:47:21 crc kubenswrapper[4873]: I1201 08:47:21.697284 4873 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/952ae5ba-156f-41a7-950a-1783e2e69f94-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:47:21 crc kubenswrapper[4873]: I1201 08:47:21.697295 4873 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/952ae5ba-156f-41a7-950a-1783e2e69f94-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 01 08:47:21 crc kubenswrapper[4873]: I1201 08:47:21.697305 4873 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/952ae5ba-156f-41a7-950a-1783e2e69f94-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 01 08:47:21 crc kubenswrapper[4873]: I1201 08:47:21.697315 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qlc6g\" (UniqueName: \"kubernetes.io/projected/952ae5ba-156f-41a7-950a-1783e2e69f94-kube-api-access-qlc6g\") on node \"crc\" DevicePath \"\"" Dec 01 08:47:21 crc kubenswrapper[4873]: I1201 08:47:21.697324 4873 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/952ae5ba-156f-41a7-950a-1783e2e69f94-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 01 08:47:21 crc kubenswrapper[4873]: I1201 08:47:21.750516 4873 generic.go:334] "Generic (PLEG): container finished" podID="952ae5ba-156f-41a7-950a-1783e2e69f94" containerID="be35e87e40c6a12a9a8bf6f80153adc4d38b8d8466027397bc51f10e7db3f1c0" exitCode=0 Dec 01 08:47:21 crc kubenswrapper[4873]: I1201 08:47:21.750601 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" event={"ID":"952ae5ba-156f-41a7-950a-1783e2e69f94","Type":"ContainerDied","Data":"be35e87e40c6a12a9a8bf6f80153adc4d38b8d8466027397bc51f10e7db3f1c0"} Dec 01 08:47:21 crc kubenswrapper[4873]: I1201 08:47:21.750636 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" Dec 01 08:47:21 crc kubenswrapper[4873]: I1201 08:47:21.750698 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-cjn5c" event={"ID":"952ae5ba-156f-41a7-950a-1783e2e69f94","Type":"ContainerDied","Data":"dc0be1a86597c49c7588c707a48263eb632a20b62d5ad5675cf32e8dfe524f9b"} Dec 01 08:47:21 crc kubenswrapper[4873]: I1201 08:47:21.750761 4873 scope.go:117] "RemoveContainer" containerID="be35e87e40c6a12a9a8bf6f80153adc4d38b8d8466027397bc51f10e7db3f1c0" Dec 01 08:47:21 crc kubenswrapper[4873]: I1201 08:47:21.780107 4873 scope.go:117] "RemoveContainer" containerID="be35e87e40c6a12a9a8bf6f80153adc4d38b8d8466027397bc51f10e7db3f1c0" Dec 01 08:47:21 crc kubenswrapper[4873]: E1201 08:47:21.781730 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be35e87e40c6a12a9a8bf6f80153adc4d38b8d8466027397bc51f10e7db3f1c0\": container with ID starting with be35e87e40c6a12a9a8bf6f80153adc4d38b8d8466027397bc51f10e7db3f1c0 not found: ID does not exist" containerID="be35e87e40c6a12a9a8bf6f80153adc4d38b8d8466027397bc51f10e7db3f1c0" Dec 01 08:47:21 crc kubenswrapper[4873]: I1201 08:47:21.781768 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be35e87e40c6a12a9a8bf6f80153adc4d38b8d8466027397bc51f10e7db3f1c0"} err="failed to get container status \"be35e87e40c6a12a9a8bf6f80153adc4d38b8d8466027397bc51f10e7db3f1c0\": rpc error: code = NotFound desc = could not find container \"be35e87e40c6a12a9a8bf6f80153adc4d38b8d8466027397bc51f10e7db3f1c0\": container with ID starting with be35e87e40c6a12a9a8bf6f80153adc4d38b8d8466027397bc51f10e7db3f1c0 not found: ID does not exist" Dec 01 08:47:21 crc kubenswrapper[4873]: I1201 08:47:21.795233 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cjn5c"] Dec 01 08:47:21 crc kubenswrapper[4873]: I1201 08:47:21.802137 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cjn5c"] Dec 01 08:47:22 crc kubenswrapper[4873]: I1201 08:47:22.442606 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="952ae5ba-156f-41a7-950a-1783e2e69f94" path="/var/lib/kubelet/pods/952ae5ba-156f-41a7-950a-1783e2e69f94/volumes" Dec 01 08:49:01 crc kubenswrapper[4873]: I1201 08:49:01.059601 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:49:01 crc kubenswrapper[4873]: I1201 08:49:01.061252 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:49:31 crc kubenswrapper[4873]: I1201 08:49:31.059497 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:49:31 crc kubenswrapper[4873]: I1201 08:49:31.061351 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:50:01 crc kubenswrapper[4873]: I1201 08:50:01.059314 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:50:01 crc kubenswrapper[4873]: I1201 08:50:01.060152 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:50:01 crc kubenswrapper[4873]: I1201 08:50:01.060227 4873 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" Dec 01 08:50:01 crc kubenswrapper[4873]: I1201 08:50:01.061197 4873 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0d13b8cddab09d2b96e4bfebd3f2cc00fd3cfd06d73474412b9d0387a8e1d7e0"} pod="openshift-machine-config-operator/machine-config-daemon-scwpp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 08:50:01 crc kubenswrapper[4873]: I1201 08:50:01.061297 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" containerID="cri-o://0d13b8cddab09d2b96e4bfebd3f2cc00fd3cfd06d73474412b9d0387a8e1d7e0" gracePeriod=600 Dec 01 08:50:01 crc kubenswrapper[4873]: I1201 08:50:01.826836 4873 generic.go:334] "Generic (PLEG): container finished" podID="fef7b114-0e07-402d-a37b-315c36011f4b" containerID="0d13b8cddab09d2b96e4bfebd3f2cc00fd3cfd06d73474412b9d0387a8e1d7e0" exitCode=0 Dec 01 08:50:01 crc kubenswrapper[4873]: I1201 08:50:01.826916 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" event={"ID":"fef7b114-0e07-402d-a37b-315c36011f4b","Type":"ContainerDied","Data":"0d13b8cddab09d2b96e4bfebd3f2cc00fd3cfd06d73474412b9d0387a8e1d7e0"} Dec 01 08:50:01 crc kubenswrapper[4873]: I1201 08:50:01.827524 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" event={"ID":"fef7b114-0e07-402d-a37b-315c36011f4b","Type":"ContainerStarted","Data":"6650a48ef92f241e9c56ee0be1ecde624a461ff74d058cf63149c0cd17ae8d2d"} Dec 01 08:50:01 crc kubenswrapper[4873]: I1201 08:50:01.827557 4873 scope.go:117] "RemoveContainer" containerID="9c30061eb3de4736a094430a40ceb3b59f45bd35deebb43c4c6aac312bafc4a8" Dec 01 08:51:10 crc kubenswrapper[4873]: I1201 08:51:10.301151 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-47tnm"] Dec 01 08:51:10 crc kubenswrapper[4873]: E1201 08:51:10.302068 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="952ae5ba-156f-41a7-950a-1783e2e69f94" containerName="registry" Dec 01 08:51:10 crc kubenswrapper[4873]: I1201 08:51:10.302085 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="952ae5ba-156f-41a7-950a-1783e2e69f94" containerName="registry" Dec 01 08:51:10 crc kubenswrapper[4873]: I1201 08:51:10.302203 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="952ae5ba-156f-41a7-950a-1783e2e69f94" containerName="registry" Dec 01 08:51:10 crc kubenswrapper[4873]: I1201 08:51:10.302810 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-47tnm" Dec 01 08:51:10 crc kubenswrapper[4873]: I1201 08:51:10.305988 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-blv72"] Dec 01 08:51:10 crc kubenswrapper[4873]: I1201 08:51:10.306099 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 01 08:51:10 crc kubenswrapper[4873]: I1201 08:51:10.306170 4873 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-8597z" Dec 01 08:51:10 crc kubenswrapper[4873]: I1201 08:51:10.306391 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 01 08:51:10 crc kubenswrapper[4873]: I1201 08:51:10.306826 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-blv72" Dec 01 08:51:10 crc kubenswrapper[4873]: I1201 08:51:10.314659 4873 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-49c9h" Dec 01 08:51:10 crc kubenswrapper[4873]: I1201 08:51:10.320221 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-blv72"] Dec 01 08:51:10 crc kubenswrapper[4873]: I1201 08:51:10.324140 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-47tnm"] Dec 01 08:51:10 crc kubenswrapper[4873]: I1201 08:51:10.327956 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-tj5cg"] Dec 01 08:51:10 crc kubenswrapper[4873]: I1201 08:51:10.328841 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-tj5cg" Dec 01 08:51:10 crc kubenswrapper[4873]: I1201 08:51:10.332272 4873 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-szs9n" Dec 01 08:51:10 crc kubenswrapper[4873]: I1201 08:51:10.344144 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-tj5cg"] Dec 01 08:51:10 crc kubenswrapper[4873]: I1201 08:51:10.377692 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgbr9\" (UniqueName: \"kubernetes.io/projected/c52cc19c-1588-49cd-a54b-072c33505a5a-kube-api-access-sgbr9\") pod \"cert-manager-5b446d88c5-blv72\" (UID: \"c52cc19c-1588-49cd-a54b-072c33505a5a\") " pod="cert-manager/cert-manager-5b446d88c5-blv72" Dec 01 08:51:10 crc kubenswrapper[4873]: I1201 08:51:10.377768 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhszh\" (UniqueName: \"kubernetes.io/projected/3876e535-8b83-41bf-8b75-2b21a25e1ff4-kube-api-access-lhszh\") pod \"cert-manager-cainjector-7f985d654d-47tnm\" (UID: \"3876e535-8b83-41bf-8b75-2b21a25e1ff4\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-47tnm" Dec 01 08:51:10 crc kubenswrapper[4873]: I1201 08:51:10.377869 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9ngq\" (UniqueName: \"kubernetes.io/projected/eb0ae03d-2543-42c9-9c93-628cab31efe2-kube-api-access-j9ngq\") pod \"cert-manager-webhook-5655c58dd6-tj5cg\" (UID: \"eb0ae03d-2543-42c9-9c93-628cab31efe2\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-tj5cg" Dec 01 08:51:10 crc kubenswrapper[4873]: I1201 08:51:10.479253 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgbr9\" (UniqueName: \"kubernetes.io/projected/c52cc19c-1588-49cd-a54b-072c33505a5a-kube-api-access-sgbr9\") pod \"cert-manager-5b446d88c5-blv72\" (UID: \"c52cc19c-1588-49cd-a54b-072c33505a5a\") " pod="cert-manager/cert-manager-5b446d88c5-blv72" Dec 01 08:51:10 crc kubenswrapper[4873]: I1201 08:51:10.479324 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhszh\" (UniqueName: \"kubernetes.io/projected/3876e535-8b83-41bf-8b75-2b21a25e1ff4-kube-api-access-lhszh\") pod \"cert-manager-cainjector-7f985d654d-47tnm\" (UID: \"3876e535-8b83-41bf-8b75-2b21a25e1ff4\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-47tnm" Dec 01 08:51:10 crc kubenswrapper[4873]: I1201 08:51:10.479361 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9ngq\" (UniqueName: \"kubernetes.io/projected/eb0ae03d-2543-42c9-9c93-628cab31efe2-kube-api-access-j9ngq\") pod \"cert-manager-webhook-5655c58dd6-tj5cg\" (UID: \"eb0ae03d-2543-42c9-9c93-628cab31efe2\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-tj5cg" Dec 01 08:51:10 crc kubenswrapper[4873]: I1201 08:51:10.501575 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgbr9\" (UniqueName: \"kubernetes.io/projected/c52cc19c-1588-49cd-a54b-072c33505a5a-kube-api-access-sgbr9\") pod \"cert-manager-5b446d88c5-blv72\" (UID: \"c52cc19c-1588-49cd-a54b-072c33505a5a\") " pod="cert-manager/cert-manager-5b446d88c5-blv72" Dec 01 08:51:10 crc kubenswrapper[4873]: I1201 08:51:10.502710 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhszh\" (UniqueName: \"kubernetes.io/projected/3876e535-8b83-41bf-8b75-2b21a25e1ff4-kube-api-access-lhszh\") pod \"cert-manager-cainjector-7f985d654d-47tnm\" (UID: \"3876e535-8b83-41bf-8b75-2b21a25e1ff4\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-47tnm" Dec 01 08:51:10 crc kubenswrapper[4873]: I1201 08:51:10.504195 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9ngq\" (UniqueName: \"kubernetes.io/projected/eb0ae03d-2543-42c9-9c93-628cab31efe2-kube-api-access-j9ngq\") pod \"cert-manager-webhook-5655c58dd6-tj5cg\" (UID: \"eb0ae03d-2543-42c9-9c93-628cab31efe2\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-tj5cg" Dec 01 08:51:10 crc kubenswrapper[4873]: I1201 08:51:10.625429 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-47tnm" Dec 01 08:51:10 crc kubenswrapper[4873]: I1201 08:51:10.639766 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-blv72" Dec 01 08:51:10 crc kubenswrapper[4873]: I1201 08:51:10.656009 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-tj5cg" Dec 01 08:51:10 crc kubenswrapper[4873]: I1201 08:51:10.909485 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-blv72"] Dec 01 08:51:10 crc kubenswrapper[4873]: I1201 08:51:10.911741 4873 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 08:51:11 crc kubenswrapper[4873]: I1201 08:51:11.148220 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-47tnm"] Dec 01 08:51:11 crc kubenswrapper[4873]: I1201 08:51:11.152896 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-tj5cg"] Dec 01 08:51:11 crc kubenswrapper[4873]: W1201 08:51:11.159730 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb0ae03d_2543_42c9_9c93_628cab31efe2.slice/crio-1b6db622a99911d839723945175126dfbbc87198e0118ba99b30fb8d1b707fe0 WatchSource:0}: Error finding container 1b6db622a99911d839723945175126dfbbc87198e0118ba99b30fb8d1b707fe0: Status 404 returned error can't find the container with id 1b6db622a99911d839723945175126dfbbc87198e0118ba99b30fb8d1b707fe0 Dec 01 08:51:11 crc kubenswrapper[4873]: I1201 08:51:11.885189 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-47tnm" event={"ID":"3876e535-8b83-41bf-8b75-2b21a25e1ff4","Type":"ContainerStarted","Data":"c40406d8e8aa709e3e5c085e9a4511a40973d795de668c956b66bdf459b64597"} Dec 01 08:51:11 crc kubenswrapper[4873]: I1201 08:51:11.886850 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-tj5cg" event={"ID":"eb0ae03d-2543-42c9-9c93-628cab31efe2","Type":"ContainerStarted","Data":"1b6db622a99911d839723945175126dfbbc87198e0118ba99b30fb8d1b707fe0"} Dec 01 08:51:11 crc kubenswrapper[4873]: I1201 08:51:11.888565 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-blv72" event={"ID":"c52cc19c-1588-49cd-a54b-072c33505a5a","Type":"ContainerStarted","Data":"fff4f474dfe6df2896068054b102c0f8a60a9ba6da1239111b73a7e9cbc73d58"} Dec 01 08:51:14 crc kubenswrapper[4873]: I1201 08:51:14.907855 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-tj5cg" event={"ID":"eb0ae03d-2543-42c9-9c93-628cab31efe2","Type":"ContainerStarted","Data":"50af92e77be4f7c7ce9050cd3a6b1649b3aecabe7f7f8ac6cd6cbe88026130a4"} Dec 01 08:51:14 crc kubenswrapper[4873]: I1201 08:51:14.908467 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-tj5cg" Dec 01 08:51:14 crc kubenswrapper[4873]: I1201 08:51:14.909980 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-blv72" event={"ID":"c52cc19c-1588-49cd-a54b-072c33505a5a","Type":"ContainerStarted","Data":"7336db6db7177d9272916a585de0a10408a7bbb30edf47d24fc8a15b7548d7b5"} Dec 01 08:51:14 crc kubenswrapper[4873]: I1201 08:51:14.911824 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-47tnm" event={"ID":"3876e535-8b83-41bf-8b75-2b21a25e1ff4","Type":"ContainerStarted","Data":"28c1aa38b6e3f767251ebe60596d7fce7fdef604ff38f5f4702df782764c8e4d"} Dec 01 08:51:14 crc kubenswrapper[4873]: I1201 08:51:14.927529 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-tj5cg" podStartSLOduration=2.125620699 podStartE2EDuration="4.927504954s" podCreationTimestamp="2025-12-01 08:51:10 +0000 UTC" firstStartedPulling="2025-12-01 08:51:11.162339329 +0000 UTC m=+647.064447878" lastFinishedPulling="2025-12-01 08:51:13.964223584 +0000 UTC m=+649.866332133" observedRunningTime="2025-12-01 08:51:14.925548782 +0000 UTC m=+650.827657321" watchObservedRunningTime="2025-12-01 08:51:14.927504954 +0000 UTC m=+650.829613503" Dec 01 08:51:14 crc kubenswrapper[4873]: I1201 08:51:14.949320 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-blv72" podStartSLOduration=1.897224627 podStartE2EDuration="4.949300234s" podCreationTimestamp="2025-12-01 08:51:10 +0000 UTC" firstStartedPulling="2025-12-01 08:51:10.911465982 +0000 UTC m=+646.813574531" lastFinishedPulling="2025-12-01 08:51:13.963541599 +0000 UTC m=+649.865650138" observedRunningTime="2025-12-01 08:51:14.944902109 +0000 UTC m=+650.847010658" watchObservedRunningTime="2025-12-01 08:51:14.949300234 +0000 UTC m=+650.851408773" Dec 01 08:51:14 crc kubenswrapper[4873]: I1201 08:51:14.963889 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-47tnm" podStartSLOduration=2.048621287 podStartE2EDuration="4.963865347s" podCreationTimestamp="2025-12-01 08:51:10 +0000 UTC" firstStartedPulling="2025-12-01 08:51:11.15413595 +0000 UTC m=+647.056244489" lastFinishedPulling="2025-12-01 08:51:14.06938001 +0000 UTC m=+649.971488549" observedRunningTime="2025-12-01 08:51:14.963180903 +0000 UTC m=+650.865289442" watchObservedRunningTime="2025-12-01 08:51:14.963865347 +0000 UTC m=+650.865973886" Dec 01 08:51:20 crc kubenswrapper[4873]: I1201 08:51:20.659548 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-tj5cg" Dec 01 08:51:20 crc kubenswrapper[4873]: I1201 08:51:20.953668 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-dxlvk"] Dec 01 08:51:20 crc kubenswrapper[4873]: I1201 08:51:20.954152 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerName="ovn-controller" containerID="cri-o://d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff" gracePeriod=30 Dec 01 08:51:20 crc kubenswrapper[4873]: I1201 08:51:20.954179 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerName="nbdb" containerID="cri-o://d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14" gracePeriod=30 Dec 01 08:51:20 crc kubenswrapper[4873]: I1201 08:51:20.954284 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a" gracePeriod=30 Dec 01 08:51:20 crc kubenswrapper[4873]: I1201 08:51:20.954282 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerName="sbdb" containerID="cri-o://3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c" gracePeriod=30 Dec 01 08:51:20 crc kubenswrapper[4873]: I1201 08:51:20.954373 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerName="kube-rbac-proxy-node" containerID="cri-o://26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c" gracePeriod=30 Dec 01 08:51:20 crc kubenswrapper[4873]: I1201 08:51:20.954389 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerName="ovn-acl-logging" containerID="cri-o://b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d" gracePeriod=30 Dec 01 08:51:20 crc kubenswrapper[4873]: I1201 08:51:20.954439 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerName="northd" containerID="cri-o://000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b" gracePeriod=30 Dec 01 08:51:20 crc kubenswrapper[4873]: I1201 08:51:20.995441 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerName="ovnkube-controller" containerID="cri-o://d4509219c44fea10d7c78878d40843169c63e790756210df070ae2c8da444cfe" gracePeriod=30 Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.298611 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dxlvk_a308045b-f54d-42ec-8fdf-5bc3e54ef363/ovnkube-controller/3.log" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.303188 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dxlvk_a308045b-f54d-42ec-8fdf-5bc3e54ef363/ovn-acl-logging/0.log" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.303642 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dxlvk_a308045b-f54d-42ec-8fdf-5bc3e54ef363/ovn-controller/0.log" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.304186 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.367103 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-jhhl4"] Dec 01 08:51:21 crc kubenswrapper[4873]: E1201 08:51:21.367367 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerName="ovnkube-controller" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.367382 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerName="ovnkube-controller" Dec 01 08:51:21 crc kubenswrapper[4873]: E1201 08:51:21.367392 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerName="kube-rbac-proxy-node" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.367399 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerName="kube-rbac-proxy-node" Dec 01 08:51:21 crc kubenswrapper[4873]: E1201 08:51:21.367412 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerName="ovn-acl-logging" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.367420 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerName="ovn-acl-logging" Dec 01 08:51:21 crc kubenswrapper[4873]: E1201 08:51:21.367430 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerName="nbdb" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.367436 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerName="nbdb" Dec 01 08:51:21 crc kubenswrapper[4873]: E1201 08:51:21.367444 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerName="northd" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.367450 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerName="northd" Dec 01 08:51:21 crc kubenswrapper[4873]: E1201 08:51:21.367460 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerName="ovn-controller" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.367466 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerName="ovn-controller" Dec 01 08:51:21 crc kubenswrapper[4873]: E1201 08:51:21.367474 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerName="ovnkube-controller" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.367479 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerName="ovnkube-controller" Dec 01 08:51:21 crc kubenswrapper[4873]: E1201 08:51:21.367487 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerName="ovnkube-controller" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.367494 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerName="ovnkube-controller" Dec 01 08:51:21 crc kubenswrapper[4873]: E1201 08:51:21.367502 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerName="ovnkube-controller" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.367508 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerName="ovnkube-controller" Dec 01 08:51:21 crc kubenswrapper[4873]: E1201 08:51:21.367519 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerName="kubecfg-setup" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.367526 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerName="kubecfg-setup" Dec 01 08:51:21 crc kubenswrapper[4873]: E1201 08:51:21.367539 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerName="sbdb" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.367545 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerName="sbdb" Dec 01 08:51:21 crc kubenswrapper[4873]: E1201 08:51:21.367553 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerName="kube-rbac-proxy-ovn-metrics" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.367559 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerName="kube-rbac-proxy-ovn-metrics" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.367657 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerName="ovnkube-controller" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.367666 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerName="ovnkube-controller" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.367674 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerName="ovnkube-controller" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.367683 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerName="kube-rbac-proxy-node" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.367689 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerName="sbdb" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.367697 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerName="ovn-acl-logging" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.367706 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerName="ovn-controller" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.367714 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerName="kube-rbac-proxy-ovn-metrics" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.367722 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerName="northd" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.367728 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerName="nbdb" Dec 01 08:51:21 crc kubenswrapper[4873]: E1201 08:51:21.367817 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerName="ovnkube-controller" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.367824 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerName="ovnkube-controller" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.367954 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerName="ovnkube-controller" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.367966 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerName="ovnkube-controller" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.369925 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.499444 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-host-run-ovn-kubernetes\") pod \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.499515 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-log-socket\") pod \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.499536 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-host-cni-netd\") pod \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.499565 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-host-cni-bin\") pod \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.499567 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-log-socket" (OuterVolumeSpecName: "log-socket") pod "a308045b-f54d-42ec-8fdf-5bc3e54ef363" (UID: "a308045b-f54d-42ec-8fdf-5bc3e54ef363"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.499569 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "a308045b-f54d-42ec-8fdf-5bc3e54ef363" (UID: "a308045b-f54d-42ec-8fdf-5bc3e54ef363"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.499603 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-systemd-units\") pod \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.499602 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "a308045b-f54d-42ec-8fdf-5bc3e54ef363" (UID: "a308045b-f54d-42ec-8fdf-5bc3e54ef363"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.499622 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "a308045b-f54d-42ec-8fdf-5bc3e54ef363" (UID: "a308045b-f54d-42ec-8fdf-5bc3e54ef363"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.499624 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "a308045b-f54d-42ec-8fdf-5bc3e54ef363" (UID: "a308045b-f54d-42ec-8fdf-5bc3e54ef363"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.499679 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a308045b-f54d-42ec-8fdf-5bc3e54ef363-ovnkube-script-lib\") pod \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.499747 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a308045b-f54d-42ec-8fdf-5bc3e54ef363-ovn-node-metrics-cert\") pod \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.500286 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a308045b-f54d-42ec-8fdf-5bc3e54ef363-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "a308045b-f54d-42ec-8fdf-5bc3e54ef363" (UID: "a308045b-f54d-42ec-8fdf-5bc3e54ef363"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.500361 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-run-ovn\") pod \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.500380 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-run-openvswitch\") pod \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.500432 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "a308045b-f54d-42ec-8fdf-5bc3e54ef363" (UID: "a308045b-f54d-42ec-8fdf-5bc3e54ef363"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.500472 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dxfd8\" (UniqueName: \"kubernetes.io/projected/a308045b-f54d-42ec-8fdf-5bc3e54ef363-kube-api-access-dxfd8\") pod \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.500497 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-var-lib-openvswitch\") pod \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.501092 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-etc-openvswitch\") pod \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.500528 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "a308045b-f54d-42ec-8fdf-5bc3e54ef363" (UID: "a308045b-f54d-42ec-8fdf-5bc3e54ef363"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.500552 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "a308045b-f54d-42ec-8fdf-5bc3e54ef363" (UID: "a308045b-f54d-42ec-8fdf-5bc3e54ef363"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.501148 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a308045b-f54d-42ec-8fdf-5bc3e54ef363-ovnkube-config\") pod \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.501195 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "a308045b-f54d-42ec-8fdf-5bc3e54ef363" (UID: "a308045b-f54d-42ec-8fdf-5bc3e54ef363"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.501205 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-host-slash\") pod \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.501225 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-host-slash" (OuterVolumeSpecName: "host-slash") pod "a308045b-f54d-42ec-8fdf-5bc3e54ef363" (UID: "a308045b-f54d-42ec-8fdf-5bc3e54ef363"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.501247 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-host-run-netns\") pod \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.501272 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-node-log\") pod \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.501298 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-host-kubelet\") pod \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.501335 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-host-var-lib-cni-networks-ovn-kubernetes\") pod \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.501362 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a308045b-f54d-42ec-8fdf-5bc3e54ef363-env-overrides\") pod \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.501383 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-run-systemd\") pod \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\" (UID: \"a308045b-f54d-42ec-8fdf-5bc3e54ef363\") " Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.501299 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-node-log" (OuterVolumeSpecName: "node-log") pod "a308045b-f54d-42ec-8fdf-5bc3e54ef363" (UID: "a308045b-f54d-42ec-8fdf-5bc3e54ef363"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.501305 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "a308045b-f54d-42ec-8fdf-5bc3e54ef363" (UID: "a308045b-f54d-42ec-8fdf-5bc3e54ef363"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.501314 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "a308045b-f54d-42ec-8fdf-5bc3e54ef363" (UID: "a308045b-f54d-42ec-8fdf-5bc3e54ef363"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.501398 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "a308045b-f54d-42ec-8fdf-5bc3e54ef363" (UID: "a308045b-f54d-42ec-8fdf-5bc3e54ef363"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.501626 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3afce38c-077a-4b3c-8f14-c799f411d883-ovnkube-config\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.501666 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3afce38c-077a-4b3c-8f14-c799f411d883-env-overrides\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.501694 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3afce38c-077a-4b3c-8f14-c799f411d883-host-kubelet\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.501702 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a308045b-f54d-42ec-8fdf-5bc3e54ef363-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "a308045b-f54d-42ec-8fdf-5bc3e54ef363" (UID: "a308045b-f54d-42ec-8fdf-5bc3e54ef363"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.501721 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3afce38c-077a-4b3c-8f14-c799f411d883-etc-openvswitch\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.501782 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3afce38c-077a-4b3c-8f14-c799f411d883-ovnkube-script-lib\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.501794 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a308045b-f54d-42ec-8fdf-5bc3e54ef363-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "a308045b-f54d-42ec-8fdf-5bc3e54ef363" (UID: "a308045b-f54d-42ec-8fdf-5bc3e54ef363"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.501807 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3afce38c-077a-4b3c-8f14-c799f411d883-host-run-ovn-kubernetes\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.501835 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3afce38c-077a-4b3c-8f14-c799f411d883-run-systemd\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.501863 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3afce38c-077a-4b3c-8f14-c799f411d883-host-cni-netd\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.501918 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3afce38c-077a-4b3c-8f14-c799f411d883-host-slash\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.501952 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3afce38c-077a-4b3c-8f14-c799f411d883-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.501987 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3afce38c-077a-4b3c-8f14-c799f411d883-node-log\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.502040 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7h2p\" (UniqueName: \"kubernetes.io/projected/3afce38c-077a-4b3c-8f14-c799f411d883-kube-api-access-m7h2p\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.502077 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3afce38c-077a-4b3c-8f14-c799f411d883-systemd-units\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.502104 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3afce38c-077a-4b3c-8f14-c799f411d883-host-cni-bin\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.502130 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3afce38c-077a-4b3c-8f14-c799f411d883-run-openvswitch\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.502148 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3afce38c-077a-4b3c-8f14-c799f411d883-var-lib-openvswitch\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.502171 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3afce38c-077a-4b3c-8f14-c799f411d883-log-socket\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.502197 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3afce38c-077a-4b3c-8f14-c799f411d883-run-ovn\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.502220 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3afce38c-077a-4b3c-8f14-c799f411d883-host-run-netns\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.502257 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3afce38c-077a-4b3c-8f14-c799f411d883-ovn-node-metrics-cert\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.502324 4873 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a308045b-f54d-42ec-8fdf-5bc3e54ef363-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.502337 4873 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-host-slash\") on node \"crc\" DevicePath \"\"" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.502346 4873 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.502359 4873 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-node-log\") on node \"crc\" DevicePath \"\"" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.502375 4873 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.502385 4873 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.502394 4873 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a308045b-f54d-42ec-8fdf-5bc3e54ef363-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.502403 4873 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.502415 4873 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-log-socket\") on node \"crc\" DevicePath \"\"" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.502424 4873 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.502432 4873 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.502440 4873 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.502450 4873 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a308045b-f54d-42ec-8fdf-5bc3e54ef363-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.502460 4873 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.502469 4873 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.502477 4873 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.502485 4873 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.506175 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a308045b-f54d-42ec-8fdf-5bc3e54ef363-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "a308045b-f54d-42ec-8fdf-5bc3e54ef363" (UID: "a308045b-f54d-42ec-8fdf-5bc3e54ef363"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.506245 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a308045b-f54d-42ec-8fdf-5bc3e54ef363-kube-api-access-dxfd8" (OuterVolumeSpecName: "kube-api-access-dxfd8") pod "a308045b-f54d-42ec-8fdf-5bc3e54ef363" (UID: "a308045b-f54d-42ec-8fdf-5bc3e54ef363"). InnerVolumeSpecName "kube-api-access-dxfd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.563164 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "a308045b-f54d-42ec-8fdf-5bc3e54ef363" (UID: "a308045b-f54d-42ec-8fdf-5bc3e54ef363"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.603959 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3afce38c-077a-4b3c-8f14-c799f411d883-host-slash\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.604093 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3afce38c-077a-4b3c-8f14-c799f411d883-node-log\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.604115 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3afce38c-077a-4b3c-8f14-c799f411d883-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.604138 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7h2p\" (UniqueName: \"kubernetes.io/projected/3afce38c-077a-4b3c-8f14-c799f411d883-kube-api-access-m7h2p\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.604129 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3afce38c-077a-4b3c-8f14-c799f411d883-host-slash\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.604248 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3afce38c-077a-4b3c-8f14-c799f411d883-node-log\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.604249 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3afce38c-077a-4b3c-8f14-c799f411d883-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.604312 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3afce38c-077a-4b3c-8f14-c799f411d883-systemd-units\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.604167 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3afce38c-077a-4b3c-8f14-c799f411d883-systemd-units\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.604373 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3afce38c-077a-4b3c-8f14-c799f411d883-host-cni-bin\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.604446 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3afce38c-077a-4b3c-8f14-c799f411d883-run-openvswitch\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.604495 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3afce38c-077a-4b3c-8f14-c799f411d883-var-lib-openvswitch\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.604522 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3afce38c-077a-4b3c-8f14-c799f411d883-host-cni-bin\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.604579 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3afce38c-077a-4b3c-8f14-c799f411d883-run-openvswitch\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.604588 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3afce38c-077a-4b3c-8f14-c799f411d883-log-socket\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.604540 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3afce38c-077a-4b3c-8f14-c799f411d883-log-socket\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.604624 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3afce38c-077a-4b3c-8f14-c799f411d883-var-lib-openvswitch\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.604647 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3afce38c-077a-4b3c-8f14-c799f411d883-run-ovn\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.604682 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3afce38c-077a-4b3c-8f14-c799f411d883-host-run-netns\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.604719 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3afce38c-077a-4b3c-8f14-c799f411d883-ovn-node-metrics-cert\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.604755 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3afce38c-077a-4b3c-8f14-c799f411d883-ovnkube-config\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.604762 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3afce38c-077a-4b3c-8f14-c799f411d883-run-ovn\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.604779 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3afce38c-077a-4b3c-8f14-c799f411d883-env-overrides\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.604782 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3afce38c-077a-4b3c-8f14-c799f411d883-host-run-netns\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.605469 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3afce38c-077a-4b3c-8f14-c799f411d883-host-kubelet\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.605510 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3afce38c-077a-4b3c-8f14-c799f411d883-etc-openvswitch\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.605548 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3afce38c-077a-4b3c-8f14-c799f411d883-ovnkube-script-lib\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.605580 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3afce38c-077a-4b3c-8f14-c799f411d883-host-run-ovn-kubernetes\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.605596 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3afce38c-077a-4b3c-8f14-c799f411d883-etc-openvswitch\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.605640 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3afce38c-077a-4b3c-8f14-c799f411d883-run-systemd\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.605604 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3afce38c-077a-4b3c-8f14-c799f411d883-run-systemd\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.605654 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3afce38c-077a-4b3c-8f14-c799f411d883-host-kubelet\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.605696 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3afce38c-077a-4b3c-8f14-c799f411d883-host-cni-netd\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.605747 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3afce38c-077a-4b3c-8f14-c799f411d883-host-run-ovn-kubernetes\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.605678 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3afce38c-077a-4b3c-8f14-c799f411d883-host-cni-netd\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.605961 4873 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a308045b-f54d-42ec-8fdf-5bc3e54ef363-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.605983 4873 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a308045b-f54d-42ec-8fdf-5bc3e54ef363-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.605998 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dxfd8\" (UniqueName: \"kubernetes.io/projected/a308045b-f54d-42ec-8fdf-5bc3e54ef363-kube-api-access-dxfd8\") on node \"crc\" DevicePath \"\"" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.606414 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3afce38c-077a-4b3c-8f14-c799f411d883-env-overrides\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.606558 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3afce38c-077a-4b3c-8f14-c799f411d883-ovnkube-config\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.606586 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3afce38c-077a-4b3c-8f14-c799f411d883-ovnkube-script-lib\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.609290 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3afce38c-077a-4b3c-8f14-c799f411d883-ovn-node-metrics-cert\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.620762 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7h2p\" (UniqueName: \"kubernetes.io/projected/3afce38c-077a-4b3c-8f14-c799f411d883-kube-api-access-m7h2p\") pod \"ovnkube-node-jhhl4\" (UID: \"3afce38c-077a-4b3c-8f14-c799f411d883\") " pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.686412 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.988585 4873 generic.go:334] "Generic (PLEG): container finished" podID="3afce38c-077a-4b3c-8f14-c799f411d883" containerID="f65598d5f41d178d4b2920617c6f13492903d6ef0a906474d0a84a2c160669b8" exitCode=0 Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.989252 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" event={"ID":"3afce38c-077a-4b3c-8f14-c799f411d883","Type":"ContainerDied","Data":"f65598d5f41d178d4b2920617c6f13492903d6ef0a906474d0a84a2c160669b8"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.989291 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" event={"ID":"3afce38c-077a-4b3c-8f14-c799f411d883","Type":"ContainerStarted","Data":"0000b597e203c5ef9d28e144594c49a0fbf321b2f417ffa2e8ac960e5a3b161f"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.993263 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dxlvk_a308045b-f54d-42ec-8fdf-5bc3e54ef363/ovnkube-controller/3.log" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.995933 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dxlvk_a308045b-f54d-42ec-8fdf-5bc3e54ef363/ovn-acl-logging/0.log" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.996506 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dxlvk_a308045b-f54d-42ec-8fdf-5bc3e54ef363/ovn-controller/0.log" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.996929 4873 generic.go:334] "Generic (PLEG): container finished" podID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerID="d4509219c44fea10d7c78878d40843169c63e790756210df070ae2c8da444cfe" exitCode=0 Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.996962 4873 generic.go:334] "Generic (PLEG): container finished" podID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerID="3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c" exitCode=0 Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.996972 4873 generic.go:334] "Generic (PLEG): container finished" podID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerID="d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14" exitCode=0 Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.996987 4873 generic.go:334] "Generic (PLEG): container finished" podID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerID="000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b" exitCode=0 Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.996999 4873 generic.go:334] "Generic (PLEG): container finished" podID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerID="61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a" exitCode=0 Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997009 4873 generic.go:334] "Generic (PLEG): container finished" podID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerID="26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c" exitCode=0 Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997032 4873 generic.go:334] "Generic (PLEG): container finished" podID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerID="b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d" exitCode=143 Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997025 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" event={"ID":"a308045b-f54d-42ec-8fdf-5bc3e54ef363","Type":"ContainerDied","Data":"d4509219c44fea10d7c78878d40843169c63e790756210df070ae2c8da444cfe"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997063 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997123 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" event={"ID":"a308045b-f54d-42ec-8fdf-5bc3e54ef363","Type":"ContainerDied","Data":"3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997152 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" event={"ID":"a308045b-f54d-42ec-8fdf-5bc3e54ef363","Type":"ContainerDied","Data":"d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997170 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" event={"ID":"a308045b-f54d-42ec-8fdf-5bc3e54ef363","Type":"ContainerDied","Data":"000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997172 4873 scope.go:117] "RemoveContainer" containerID="d4509219c44fea10d7c78878d40843169c63e790756210df070ae2c8da444cfe" Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997183 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" event={"ID":"a308045b-f54d-42ec-8fdf-5bc3e54ef363","Type":"ContainerDied","Data":"61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997196 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" event={"ID":"a308045b-f54d-42ec-8fdf-5bc3e54ef363","Type":"ContainerDied","Data":"26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997210 4873 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c55ab10131e65cc646cead8704fd95f3534c6726b01283d2e3fdf4b8da4e03ff"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997226 4873 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997234 4873 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997244 4873 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997044 4873 generic.go:334] "Generic (PLEG): container finished" podID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" containerID="d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff" exitCode=143 Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997251 4873 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997391 4873 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997427 4873 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997437 4873 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997448 4873 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997479 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" event={"ID":"a308045b-f54d-42ec-8fdf-5bc3e54ef363","Type":"ContainerDied","Data":"b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997522 4873 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d4509219c44fea10d7c78878d40843169c63e790756210df070ae2c8da444cfe"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997531 4873 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c55ab10131e65cc646cead8704fd95f3534c6726b01283d2e3fdf4b8da4e03ff"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997538 4873 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997546 4873 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997554 4873 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997561 4873 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997568 4873 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997576 4873 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997583 4873 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997591 4873 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997603 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" event={"ID":"a308045b-f54d-42ec-8fdf-5bc3e54ef363","Type":"ContainerDied","Data":"d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997617 4873 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d4509219c44fea10d7c78878d40843169c63e790756210df070ae2c8da444cfe"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997629 4873 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c55ab10131e65cc646cead8704fd95f3534c6726b01283d2e3fdf4b8da4e03ff"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997636 4873 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997643 4873 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997650 4873 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997657 4873 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997665 4873 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997674 4873 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997682 4873 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997689 4873 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997700 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dxlvk" event={"ID":"a308045b-f54d-42ec-8fdf-5bc3e54ef363","Type":"ContainerDied","Data":"c0897dd565e0e5ec410cf9ace85340d1c2f155a4c498b0a8d0d25ace45fa5f65"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997713 4873 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d4509219c44fea10d7c78878d40843169c63e790756210df070ae2c8da444cfe"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997723 4873 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c55ab10131e65cc646cead8704fd95f3534c6726b01283d2e3fdf4b8da4e03ff"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997731 4873 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997738 4873 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997745 4873 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997759 4873 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997767 4873 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997774 4873 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997780 4873 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.997787 4873 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70"} Dec 01 08:51:21 crc kubenswrapper[4873]: I1201 08:51:21.999363 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-tpwwr_13a2246b-93bb-4586-98ee-53fc84aaae02/kube-multus/2.log" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.020440 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-tpwwr_13a2246b-93bb-4586-98ee-53fc84aaae02/kube-multus/1.log" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.020524 4873 generic.go:334] "Generic (PLEG): container finished" podID="13a2246b-93bb-4586-98ee-53fc84aaae02" containerID="b80d1c593405c5217176f034e50359d6fea16f79b5f8758b7753809b6ab96f76" exitCode=2 Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.020575 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-tpwwr" event={"ID":"13a2246b-93bb-4586-98ee-53fc84aaae02","Type":"ContainerDied","Data":"b80d1c593405c5217176f034e50359d6fea16f79b5f8758b7753809b6ab96f76"} Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.020619 4873 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a9f46a2651e067cd11996fb1c43221a21ab71fdc3ca78a61d9245516f871a5f3"} Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.021437 4873 scope.go:117] "RemoveContainer" containerID="b80d1c593405c5217176f034e50359d6fea16f79b5f8758b7753809b6ab96f76" Dec 01 08:51:22 crc kubenswrapper[4873]: E1201 08:51:22.021690 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-tpwwr_openshift-multus(13a2246b-93bb-4586-98ee-53fc84aaae02)\"" pod="openshift-multus/multus-tpwwr" podUID="13a2246b-93bb-4586-98ee-53fc84aaae02" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.048204 4873 scope.go:117] "RemoveContainer" containerID="c55ab10131e65cc646cead8704fd95f3534c6726b01283d2e3fdf4b8da4e03ff" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.057500 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-dxlvk"] Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.068587 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-dxlvk"] Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.077854 4873 scope.go:117] "RemoveContainer" containerID="3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.131627 4873 scope.go:117] "RemoveContainer" containerID="d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.149558 4873 scope.go:117] "RemoveContainer" containerID="000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.166216 4873 scope.go:117] "RemoveContainer" containerID="61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.183632 4873 scope.go:117] "RemoveContainer" containerID="26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.200563 4873 scope.go:117] "RemoveContainer" containerID="b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.226890 4873 scope.go:117] "RemoveContainer" containerID="d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.258226 4873 scope.go:117] "RemoveContainer" containerID="a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.276555 4873 scope.go:117] "RemoveContainer" containerID="d4509219c44fea10d7c78878d40843169c63e790756210df070ae2c8da444cfe" Dec 01 08:51:22 crc kubenswrapper[4873]: E1201 08:51:22.277182 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4509219c44fea10d7c78878d40843169c63e790756210df070ae2c8da444cfe\": container with ID starting with d4509219c44fea10d7c78878d40843169c63e790756210df070ae2c8da444cfe not found: ID does not exist" containerID="d4509219c44fea10d7c78878d40843169c63e790756210df070ae2c8da444cfe" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.277258 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4509219c44fea10d7c78878d40843169c63e790756210df070ae2c8da444cfe"} err="failed to get container status \"d4509219c44fea10d7c78878d40843169c63e790756210df070ae2c8da444cfe\": rpc error: code = NotFound desc = could not find container \"d4509219c44fea10d7c78878d40843169c63e790756210df070ae2c8da444cfe\": container with ID starting with d4509219c44fea10d7c78878d40843169c63e790756210df070ae2c8da444cfe not found: ID does not exist" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.277300 4873 scope.go:117] "RemoveContainer" containerID="c55ab10131e65cc646cead8704fd95f3534c6726b01283d2e3fdf4b8da4e03ff" Dec 01 08:51:22 crc kubenswrapper[4873]: E1201 08:51:22.277739 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c55ab10131e65cc646cead8704fd95f3534c6726b01283d2e3fdf4b8da4e03ff\": container with ID starting with c55ab10131e65cc646cead8704fd95f3534c6726b01283d2e3fdf4b8da4e03ff not found: ID does not exist" containerID="c55ab10131e65cc646cead8704fd95f3534c6726b01283d2e3fdf4b8da4e03ff" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.277784 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c55ab10131e65cc646cead8704fd95f3534c6726b01283d2e3fdf4b8da4e03ff"} err="failed to get container status \"c55ab10131e65cc646cead8704fd95f3534c6726b01283d2e3fdf4b8da4e03ff\": rpc error: code = NotFound desc = could not find container \"c55ab10131e65cc646cead8704fd95f3534c6726b01283d2e3fdf4b8da4e03ff\": container with ID starting with c55ab10131e65cc646cead8704fd95f3534c6726b01283d2e3fdf4b8da4e03ff not found: ID does not exist" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.277814 4873 scope.go:117] "RemoveContainer" containerID="3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c" Dec 01 08:51:22 crc kubenswrapper[4873]: E1201 08:51:22.278243 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c\": container with ID starting with 3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c not found: ID does not exist" containerID="3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.278430 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c"} err="failed to get container status \"3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c\": rpc error: code = NotFound desc = could not find container \"3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c\": container with ID starting with 3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c not found: ID does not exist" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.278489 4873 scope.go:117] "RemoveContainer" containerID="d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14" Dec 01 08:51:22 crc kubenswrapper[4873]: E1201 08:51:22.278899 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14\": container with ID starting with d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14 not found: ID does not exist" containerID="d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.278931 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14"} err="failed to get container status \"d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14\": rpc error: code = NotFound desc = could not find container \"d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14\": container with ID starting with d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14 not found: ID does not exist" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.278954 4873 scope.go:117] "RemoveContainer" containerID="000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b" Dec 01 08:51:22 crc kubenswrapper[4873]: E1201 08:51:22.279223 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b\": container with ID starting with 000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b not found: ID does not exist" containerID="000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.279258 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b"} err="failed to get container status \"000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b\": rpc error: code = NotFound desc = could not find container \"000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b\": container with ID starting with 000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b not found: ID does not exist" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.279279 4873 scope.go:117] "RemoveContainer" containerID="61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a" Dec 01 08:51:22 crc kubenswrapper[4873]: E1201 08:51:22.279522 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a\": container with ID starting with 61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a not found: ID does not exist" containerID="61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.279547 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a"} err="failed to get container status \"61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a\": rpc error: code = NotFound desc = could not find container \"61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a\": container with ID starting with 61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a not found: ID does not exist" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.279561 4873 scope.go:117] "RemoveContainer" containerID="26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c" Dec 01 08:51:22 crc kubenswrapper[4873]: E1201 08:51:22.279851 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c\": container with ID starting with 26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c not found: ID does not exist" containerID="26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.279886 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c"} err="failed to get container status \"26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c\": rpc error: code = NotFound desc = could not find container \"26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c\": container with ID starting with 26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c not found: ID does not exist" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.279905 4873 scope.go:117] "RemoveContainer" containerID="b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d" Dec 01 08:51:22 crc kubenswrapper[4873]: E1201 08:51:22.280236 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d\": container with ID starting with b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d not found: ID does not exist" containerID="b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.280271 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d"} err="failed to get container status \"b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d\": rpc error: code = NotFound desc = could not find container \"b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d\": container with ID starting with b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d not found: ID does not exist" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.280294 4873 scope.go:117] "RemoveContainer" containerID="d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff" Dec 01 08:51:22 crc kubenswrapper[4873]: E1201 08:51:22.280585 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff\": container with ID starting with d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff not found: ID does not exist" containerID="d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.280619 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff"} err="failed to get container status \"d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff\": rpc error: code = NotFound desc = could not find container \"d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff\": container with ID starting with d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff not found: ID does not exist" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.280639 4873 scope.go:117] "RemoveContainer" containerID="a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70" Dec 01 08:51:22 crc kubenswrapper[4873]: E1201 08:51:22.280946 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\": container with ID starting with a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70 not found: ID does not exist" containerID="a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.280983 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70"} err="failed to get container status \"a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\": rpc error: code = NotFound desc = could not find container \"a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\": container with ID starting with a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70 not found: ID does not exist" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.281008 4873 scope.go:117] "RemoveContainer" containerID="d4509219c44fea10d7c78878d40843169c63e790756210df070ae2c8da444cfe" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.281371 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4509219c44fea10d7c78878d40843169c63e790756210df070ae2c8da444cfe"} err="failed to get container status \"d4509219c44fea10d7c78878d40843169c63e790756210df070ae2c8da444cfe\": rpc error: code = NotFound desc = could not find container \"d4509219c44fea10d7c78878d40843169c63e790756210df070ae2c8da444cfe\": container with ID starting with d4509219c44fea10d7c78878d40843169c63e790756210df070ae2c8da444cfe not found: ID does not exist" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.281402 4873 scope.go:117] "RemoveContainer" containerID="c55ab10131e65cc646cead8704fd95f3534c6726b01283d2e3fdf4b8da4e03ff" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.281708 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c55ab10131e65cc646cead8704fd95f3534c6726b01283d2e3fdf4b8da4e03ff"} err="failed to get container status \"c55ab10131e65cc646cead8704fd95f3534c6726b01283d2e3fdf4b8da4e03ff\": rpc error: code = NotFound desc = could not find container \"c55ab10131e65cc646cead8704fd95f3534c6726b01283d2e3fdf4b8da4e03ff\": container with ID starting with c55ab10131e65cc646cead8704fd95f3534c6726b01283d2e3fdf4b8da4e03ff not found: ID does not exist" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.281732 4873 scope.go:117] "RemoveContainer" containerID="3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.282024 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c"} err="failed to get container status \"3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c\": rpc error: code = NotFound desc = could not find container \"3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c\": container with ID starting with 3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c not found: ID does not exist" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.282048 4873 scope.go:117] "RemoveContainer" containerID="d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.282381 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14"} err="failed to get container status \"d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14\": rpc error: code = NotFound desc = could not find container \"d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14\": container with ID starting with d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14 not found: ID does not exist" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.282403 4873 scope.go:117] "RemoveContainer" containerID="000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.282701 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b"} err="failed to get container status \"000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b\": rpc error: code = NotFound desc = could not find container \"000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b\": container with ID starting with 000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b not found: ID does not exist" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.282727 4873 scope.go:117] "RemoveContainer" containerID="61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.283048 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a"} err="failed to get container status \"61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a\": rpc error: code = NotFound desc = could not find container \"61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a\": container with ID starting with 61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a not found: ID does not exist" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.283074 4873 scope.go:117] "RemoveContainer" containerID="26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.283461 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c"} err="failed to get container status \"26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c\": rpc error: code = NotFound desc = could not find container \"26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c\": container with ID starting with 26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c not found: ID does not exist" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.283488 4873 scope.go:117] "RemoveContainer" containerID="b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.283786 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d"} err="failed to get container status \"b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d\": rpc error: code = NotFound desc = could not find container \"b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d\": container with ID starting with b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d not found: ID does not exist" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.283810 4873 scope.go:117] "RemoveContainer" containerID="d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.284122 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff"} err="failed to get container status \"d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff\": rpc error: code = NotFound desc = could not find container \"d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff\": container with ID starting with d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff not found: ID does not exist" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.284143 4873 scope.go:117] "RemoveContainer" containerID="a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.284465 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70"} err="failed to get container status \"a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\": rpc error: code = NotFound desc = could not find container \"a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\": container with ID starting with a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70 not found: ID does not exist" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.284494 4873 scope.go:117] "RemoveContainer" containerID="d4509219c44fea10d7c78878d40843169c63e790756210df070ae2c8da444cfe" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.284951 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4509219c44fea10d7c78878d40843169c63e790756210df070ae2c8da444cfe"} err="failed to get container status \"d4509219c44fea10d7c78878d40843169c63e790756210df070ae2c8da444cfe\": rpc error: code = NotFound desc = could not find container \"d4509219c44fea10d7c78878d40843169c63e790756210df070ae2c8da444cfe\": container with ID starting with d4509219c44fea10d7c78878d40843169c63e790756210df070ae2c8da444cfe not found: ID does not exist" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.284972 4873 scope.go:117] "RemoveContainer" containerID="c55ab10131e65cc646cead8704fd95f3534c6726b01283d2e3fdf4b8da4e03ff" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.285387 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c55ab10131e65cc646cead8704fd95f3534c6726b01283d2e3fdf4b8da4e03ff"} err="failed to get container status \"c55ab10131e65cc646cead8704fd95f3534c6726b01283d2e3fdf4b8da4e03ff\": rpc error: code = NotFound desc = could not find container \"c55ab10131e65cc646cead8704fd95f3534c6726b01283d2e3fdf4b8da4e03ff\": container with ID starting with c55ab10131e65cc646cead8704fd95f3534c6726b01283d2e3fdf4b8da4e03ff not found: ID does not exist" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.285405 4873 scope.go:117] "RemoveContainer" containerID="3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.285648 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c"} err="failed to get container status \"3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c\": rpc error: code = NotFound desc = could not find container \"3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c\": container with ID starting with 3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c not found: ID does not exist" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.285673 4873 scope.go:117] "RemoveContainer" containerID="d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.287004 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14"} err="failed to get container status \"d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14\": rpc error: code = NotFound desc = could not find container \"d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14\": container with ID starting with d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14 not found: ID does not exist" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.287049 4873 scope.go:117] "RemoveContainer" containerID="000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.287396 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b"} err="failed to get container status \"000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b\": rpc error: code = NotFound desc = could not find container \"000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b\": container with ID starting with 000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b not found: ID does not exist" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.287432 4873 scope.go:117] "RemoveContainer" containerID="61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.288131 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a"} err="failed to get container status \"61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a\": rpc error: code = NotFound desc = could not find container \"61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a\": container with ID starting with 61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a not found: ID does not exist" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.288190 4873 scope.go:117] "RemoveContainer" containerID="26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.288533 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c"} err="failed to get container status \"26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c\": rpc error: code = NotFound desc = could not find container \"26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c\": container with ID starting with 26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c not found: ID does not exist" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.288558 4873 scope.go:117] "RemoveContainer" containerID="b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.288917 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d"} err="failed to get container status \"b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d\": rpc error: code = NotFound desc = could not find container \"b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d\": container with ID starting with b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d not found: ID does not exist" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.288939 4873 scope.go:117] "RemoveContainer" containerID="d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.289508 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff"} err="failed to get container status \"d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff\": rpc error: code = NotFound desc = could not find container \"d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff\": container with ID starting with d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff not found: ID does not exist" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.289545 4873 scope.go:117] "RemoveContainer" containerID="a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.289894 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70"} err="failed to get container status \"a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\": rpc error: code = NotFound desc = could not find container \"a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\": container with ID starting with a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70 not found: ID does not exist" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.289915 4873 scope.go:117] "RemoveContainer" containerID="d4509219c44fea10d7c78878d40843169c63e790756210df070ae2c8da444cfe" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.290228 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4509219c44fea10d7c78878d40843169c63e790756210df070ae2c8da444cfe"} err="failed to get container status \"d4509219c44fea10d7c78878d40843169c63e790756210df070ae2c8da444cfe\": rpc error: code = NotFound desc = could not find container \"d4509219c44fea10d7c78878d40843169c63e790756210df070ae2c8da444cfe\": container with ID starting with d4509219c44fea10d7c78878d40843169c63e790756210df070ae2c8da444cfe not found: ID does not exist" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.290259 4873 scope.go:117] "RemoveContainer" containerID="c55ab10131e65cc646cead8704fd95f3534c6726b01283d2e3fdf4b8da4e03ff" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.290517 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c55ab10131e65cc646cead8704fd95f3534c6726b01283d2e3fdf4b8da4e03ff"} err="failed to get container status \"c55ab10131e65cc646cead8704fd95f3534c6726b01283d2e3fdf4b8da4e03ff\": rpc error: code = NotFound desc = could not find container \"c55ab10131e65cc646cead8704fd95f3534c6726b01283d2e3fdf4b8da4e03ff\": container with ID starting with c55ab10131e65cc646cead8704fd95f3534c6726b01283d2e3fdf4b8da4e03ff not found: ID does not exist" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.290540 4873 scope.go:117] "RemoveContainer" containerID="3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.290852 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c"} err="failed to get container status \"3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c\": rpc error: code = NotFound desc = could not find container \"3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c\": container with ID starting with 3183b0a6669b74e4cf53edb7296855eda85f1291074064162caee3aeabffa67c not found: ID does not exist" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.290871 4873 scope.go:117] "RemoveContainer" containerID="d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.291312 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14"} err="failed to get container status \"d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14\": rpc error: code = NotFound desc = could not find container \"d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14\": container with ID starting with d7a7cb8945854b72b99fb44879aa11cd4340137c8f16c7c3a510611971860e14 not found: ID does not exist" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.291346 4873 scope.go:117] "RemoveContainer" containerID="000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.291662 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b"} err="failed to get container status \"000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b\": rpc error: code = NotFound desc = could not find container \"000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b\": container with ID starting with 000bcf3d9ff0781d7039b2945f451f0caddd7af2c36d18e82cf238058cdcdc2b not found: ID does not exist" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.291684 4873 scope.go:117] "RemoveContainer" containerID="61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.291978 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a"} err="failed to get container status \"61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a\": rpc error: code = NotFound desc = could not find container \"61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a\": container with ID starting with 61b6ccade52c73bdf73733ab5aba6306c177ceeed38aaa0640605a8ba9fe722a not found: ID does not exist" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.292008 4873 scope.go:117] "RemoveContainer" containerID="26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.292329 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c"} err="failed to get container status \"26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c\": rpc error: code = NotFound desc = could not find container \"26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c\": container with ID starting with 26610df724f0b4b4f7f6a07666b53ba3ffb122230a1797549ee4f469d0628b7c not found: ID does not exist" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.292354 4873 scope.go:117] "RemoveContainer" containerID="b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.292658 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d"} err="failed to get container status \"b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d\": rpc error: code = NotFound desc = could not find container \"b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d\": container with ID starting with b9e8cc0a06a666d8307097cd95b8227e295a4840f8958cb189ab542ce08e887d not found: ID does not exist" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.292681 4873 scope.go:117] "RemoveContainer" containerID="d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.294005 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff"} err="failed to get container status \"d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff\": rpc error: code = NotFound desc = could not find container \"d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff\": container with ID starting with d98814b7943ccadb8a254f6889e8ca5a94cf52e56f7eabc34ff14623bf2330ff not found: ID does not exist" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.294036 4873 scope.go:117] "RemoveContainer" containerID="a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.294359 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70"} err="failed to get container status \"a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\": rpc error: code = NotFound desc = could not find container \"a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70\": container with ID starting with a44a7858f03ea58c8baab78a0d84593530d673f6344e6e917df4222d9f98bb70 not found: ID does not exist" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.294394 4873 scope.go:117] "RemoveContainer" containerID="d4509219c44fea10d7c78878d40843169c63e790756210df070ae2c8da444cfe" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.294697 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4509219c44fea10d7c78878d40843169c63e790756210df070ae2c8da444cfe"} err="failed to get container status \"d4509219c44fea10d7c78878d40843169c63e790756210df070ae2c8da444cfe\": rpc error: code = NotFound desc = could not find container \"d4509219c44fea10d7c78878d40843169c63e790756210df070ae2c8da444cfe\": container with ID starting with d4509219c44fea10d7c78878d40843169c63e790756210df070ae2c8da444cfe not found: ID does not exist" Dec 01 08:51:22 crc kubenswrapper[4873]: I1201 08:51:22.437331 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a308045b-f54d-42ec-8fdf-5bc3e54ef363" path="/var/lib/kubelet/pods/a308045b-f54d-42ec-8fdf-5bc3e54ef363/volumes" Dec 01 08:51:23 crc kubenswrapper[4873]: I1201 08:51:23.030651 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" event={"ID":"3afce38c-077a-4b3c-8f14-c799f411d883","Type":"ContainerStarted","Data":"74411082b0aed51355e50439b7be86e88670213d2714c93917840d46594caac8"} Dec 01 08:51:23 crc kubenswrapper[4873]: I1201 08:51:23.031141 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" event={"ID":"3afce38c-077a-4b3c-8f14-c799f411d883","Type":"ContainerStarted","Data":"a7e780b83f32cfb247d3ed86d2f83c532a1b1c9824f130503b793cc9679b3a11"} Dec 01 08:51:23 crc kubenswrapper[4873]: I1201 08:51:23.031164 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" event={"ID":"3afce38c-077a-4b3c-8f14-c799f411d883","Type":"ContainerStarted","Data":"81a9c89fb2b917d5f6cb1ea093c8cfc0aa46e8bda4f0840af8a070772215124e"} Dec 01 08:51:23 crc kubenswrapper[4873]: I1201 08:51:23.031181 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" event={"ID":"3afce38c-077a-4b3c-8f14-c799f411d883","Type":"ContainerStarted","Data":"fd573ba5e7676126e1cafeae91029617d155dce4e59d952cd91c31a838508ef7"} Dec 01 08:51:23 crc kubenswrapper[4873]: I1201 08:51:23.031193 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" event={"ID":"3afce38c-077a-4b3c-8f14-c799f411d883","Type":"ContainerStarted","Data":"6ac0a32d6586b31aace9a1b68eba988280d1411d7d95434f7c03f21bf23b6623"} Dec 01 08:51:23 crc kubenswrapper[4873]: I1201 08:51:23.031207 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" event={"ID":"3afce38c-077a-4b3c-8f14-c799f411d883","Type":"ContainerStarted","Data":"533311aff71d297c1c2930e41d72470d360d653f571bd0b8eed233114fd538ff"} Dec 01 08:51:24 crc kubenswrapper[4873]: I1201 08:51:24.688238 4873 scope.go:117] "RemoveContainer" containerID="a9f46a2651e067cd11996fb1c43221a21ab71fdc3ca78a61d9245516f871a5f3" Dec 01 08:51:25 crc kubenswrapper[4873]: I1201 08:51:25.046626 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-tpwwr_13a2246b-93bb-4586-98ee-53fc84aaae02/kube-multus/2.log" Dec 01 08:51:26 crc kubenswrapper[4873]: I1201 08:51:26.057471 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" event={"ID":"3afce38c-077a-4b3c-8f14-c799f411d883","Type":"ContainerStarted","Data":"3019e13b73fddd91aeafd91d3b50c24152524978e75a98ffd9f4ae895e8b2e07"} Dec 01 08:51:28 crc kubenswrapper[4873]: I1201 08:51:28.074665 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" event={"ID":"3afce38c-077a-4b3c-8f14-c799f411d883","Type":"ContainerStarted","Data":"37130a6bfa785c15293b5ec5b9f06a2104de43469d14eaf2bb0f6f3c592db133"} Dec 01 08:51:28 crc kubenswrapper[4873]: I1201 08:51:28.076882 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:28 crc kubenswrapper[4873]: I1201 08:51:28.076921 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:28 crc kubenswrapper[4873]: I1201 08:51:28.076931 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:28 crc kubenswrapper[4873]: I1201 08:51:28.115769 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:28 crc kubenswrapper[4873]: I1201 08:51:28.119814 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" podStartSLOduration=7.119792852 podStartE2EDuration="7.119792852s" podCreationTimestamp="2025-12-01 08:51:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:51:28.118225188 +0000 UTC m=+664.020333737" watchObservedRunningTime="2025-12-01 08:51:28.119792852 +0000 UTC m=+664.021901391" Dec 01 08:51:28 crc kubenswrapper[4873]: I1201 08:51:28.128075 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:36 crc kubenswrapper[4873]: I1201 08:51:36.431239 4873 scope.go:117] "RemoveContainer" containerID="b80d1c593405c5217176f034e50359d6fea16f79b5f8758b7753809b6ab96f76" Dec 01 08:51:36 crc kubenswrapper[4873]: E1201 08:51:36.432263 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-tpwwr_openshift-multus(13a2246b-93bb-4586-98ee-53fc84aaae02)\"" pod="openshift-multus/multus-tpwwr" podUID="13a2246b-93bb-4586-98ee-53fc84aaae02" Dec 01 08:51:49 crc kubenswrapper[4873]: I1201 08:51:49.430176 4873 scope.go:117] "RemoveContainer" containerID="b80d1c593405c5217176f034e50359d6fea16f79b5f8758b7753809b6ab96f76" Dec 01 08:51:50 crc kubenswrapper[4873]: I1201 08:51:50.216703 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-tpwwr_13a2246b-93bb-4586-98ee-53fc84aaae02/kube-multus/2.log" Dec 01 08:51:50 crc kubenswrapper[4873]: I1201 08:51:50.217546 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-tpwwr" event={"ID":"13a2246b-93bb-4586-98ee-53fc84aaae02","Type":"ContainerStarted","Data":"5490e6161ce89d752b559c72987c004502bb301b83bad7b4d2d9ce85d74ef80b"} Dec 01 08:51:51 crc kubenswrapper[4873]: I1201 08:51:51.715591 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-jhhl4" Dec 01 08:51:59 crc kubenswrapper[4873]: I1201 08:51:59.230802 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhgjjb"] Dec 01 08:51:59 crc kubenswrapper[4873]: I1201 08:51:59.233057 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhgjjb" Dec 01 08:51:59 crc kubenswrapper[4873]: I1201 08:51:59.237697 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 01 08:51:59 crc kubenswrapper[4873]: I1201 08:51:59.241350 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhgjjb"] Dec 01 08:51:59 crc kubenswrapper[4873]: I1201 08:51:59.288478 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b77a8d6f-5a83-494e-a366-300ffde2afb4-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhgjjb\" (UID: \"b77a8d6f-5a83-494e-a366-300ffde2afb4\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhgjjb" Dec 01 08:51:59 crc kubenswrapper[4873]: I1201 08:51:59.288599 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pdnf\" (UniqueName: \"kubernetes.io/projected/b77a8d6f-5a83-494e-a366-300ffde2afb4-kube-api-access-5pdnf\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhgjjb\" (UID: \"b77a8d6f-5a83-494e-a366-300ffde2afb4\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhgjjb" Dec 01 08:51:59 crc kubenswrapper[4873]: I1201 08:51:59.288749 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b77a8d6f-5a83-494e-a366-300ffde2afb4-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhgjjb\" (UID: \"b77a8d6f-5a83-494e-a366-300ffde2afb4\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhgjjb" Dec 01 08:51:59 crc kubenswrapper[4873]: I1201 08:51:59.389538 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b77a8d6f-5a83-494e-a366-300ffde2afb4-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhgjjb\" (UID: \"b77a8d6f-5a83-494e-a366-300ffde2afb4\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhgjjb" Dec 01 08:51:59 crc kubenswrapper[4873]: I1201 08:51:59.389610 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b77a8d6f-5a83-494e-a366-300ffde2afb4-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhgjjb\" (UID: \"b77a8d6f-5a83-494e-a366-300ffde2afb4\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhgjjb" Dec 01 08:51:59 crc kubenswrapper[4873]: I1201 08:51:59.389701 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pdnf\" (UniqueName: \"kubernetes.io/projected/b77a8d6f-5a83-494e-a366-300ffde2afb4-kube-api-access-5pdnf\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhgjjb\" (UID: \"b77a8d6f-5a83-494e-a366-300ffde2afb4\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhgjjb" Dec 01 08:51:59 crc kubenswrapper[4873]: I1201 08:51:59.390231 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b77a8d6f-5a83-494e-a366-300ffde2afb4-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhgjjb\" (UID: \"b77a8d6f-5a83-494e-a366-300ffde2afb4\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhgjjb" Dec 01 08:51:59 crc kubenswrapper[4873]: I1201 08:51:59.390285 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b77a8d6f-5a83-494e-a366-300ffde2afb4-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhgjjb\" (UID: \"b77a8d6f-5a83-494e-a366-300ffde2afb4\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhgjjb" Dec 01 08:51:59 crc kubenswrapper[4873]: I1201 08:51:59.408707 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pdnf\" (UniqueName: \"kubernetes.io/projected/b77a8d6f-5a83-494e-a366-300ffde2afb4-kube-api-access-5pdnf\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhgjjb\" (UID: \"b77a8d6f-5a83-494e-a366-300ffde2afb4\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhgjjb" Dec 01 08:51:59 crc kubenswrapper[4873]: I1201 08:51:59.551785 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhgjjb" Dec 01 08:51:59 crc kubenswrapper[4873]: I1201 08:51:59.787450 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhgjjb"] Dec 01 08:52:00 crc kubenswrapper[4873]: I1201 08:52:00.281902 4873 generic.go:334] "Generic (PLEG): container finished" podID="b77a8d6f-5a83-494e-a366-300ffde2afb4" containerID="e7c49e834e9633a064849e722db56c8fcad0c0a88e3d56ec7b21a533e1fc0394" exitCode=0 Dec 01 08:52:00 crc kubenswrapper[4873]: I1201 08:52:00.281958 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhgjjb" event={"ID":"b77a8d6f-5a83-494e-a366-300ffde2afb4","Type":"ContainerDied","Data":"e7c49e834e9633a064849e722db56c8fcad0c0a88e3d56ec7b21a533e1fc0394"} Dec 01 08:52:00 crc kubenswrapper[4873]: I1201 08:52:00.281998 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhgjjb" event={"ID":"b77a8d6f-5a83-494e-a366-300ffde2afb4","Type":"ContainerStarted","Data":"9129afb0477cab9f7df107dfa0be8546394ea628f920f598c1b30ffe113cf39c"} Dec 01 08:52:01 crc kubenswrapper[4873]: I1201 08:52:01.059587 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:52:01 crc kubenswrapper[4873]: I1201 08:52:01.060158 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:52:02 crc kubenswrapper[4873]: I1201 08:52:02.315459 4873 generic.go:334] "Generic (PLEG): container finished" podID="b77a8d6f-5a83-494e-a366-300ffde2afb4" containerID="7580c52f1a81b16cb138ccfc51c08b791ba72df915510a359225d9531b1eac87" exitCode=0 Dec 01 08:52:02 crc kubenswrapper[4873]: I1201 08:52:02.315535 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhgjjb" event={"ID":"b77a8d6f-5a83-494e-a366-300ffde2afb4","Type":"ContainerDied","Data":"7580c52f1a81b16cb138ccfc51c08b791ba72df915510a359225d9531b1eac87"} Dec 01 08:52:03 crc kubenswrapper[4873]: I1201 08:52:03.326383 4873 generic.go:334] "Generic (PLEG): container finished" podID="b77a8d6f-5a83-494e-a366-300ffde2afb4" containerID="d2911c4c3d8e2e22669afb7daa1dcf67531e85f072c7dc26f3e19b2c57c95362" exitCode=0 Dec 01 08:52:03 crc kubenswrapper[4873]: I1201 08:52:03.326490 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhgjjb" event={"ID":"b77a8d6f-5a83-494e-a366-300ffde2afb4","Type":"ContainerDied","Data":"d2911c4c3d8e2e22669afb7daa1dcf67531e85f072c7dc26f3e19b2c57c95362"} Dec 01 08:52:04 crc kubenswrapper[4873]: I1201 08:52:04.554133 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhgjjb" Dec 01 08:52:04 crc kubenswrapper[4873]: I1201 08:52:04.672722 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5pdnf\" (UniqueName: \"kubernetes.io/projected/b77a8d6f-5a83-494e-a366-300ffde2afb4-kube-api-access-5pdnf\") pod \"b77a8d6f-5a83-494e-a366-300ffde2afb4\" (UID: \"b77a8d6f-5a83-494e-a366-300ffde2afb4\") " Dec 01 08:52:04 crc kubenswrapper[4873]: I1201 08:52:04.672806 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b77a8d6f-5a83-494e-a366-300ffde2afb4-util\") pod \"b77a8d6f-5a83-494e-a366-300ffde2afb4\" (UID: \"b77a8d6f-5a83-494e-a366-300ffde2afb4\") " Dec 01 08:52:04 crc kubenswrapper[4873]: I1201 08:52:04.672837 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b77a8d6f-5a83-494e-a366-300ffde2afb4-bundle\") pod \"b77a8d6f-5a83-494e-a366-300ffde2afb4\" (UID: \"b77a8d6f-5a83-494e-a366-300ffde2afb4\") " Dec 01 08:52:04 crc kubenswrapper[4873]: I1201 08:52:04.673727 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b77a8d6f-5a83-494e-a366-300ffde2afb4-bundle" (OuterVolumeSpecName: "bundle") pod "b77a8d6f-5a83-494e-a366-300ffde2afb4" (UID: "b77a8d6f-5a83-494e-a366-300ffde2afb4"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:52:04 crc kubenswrapper[4873]: I1201 08:52:04.678732 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b77a8d6f-5a83-494e-a366-300ffde2afb4-kube-api-access-5pdnf" (OuterVolumeSpecName: "kube-api-access-5pdnf") pod "b77a8d6f-5a83-494e-a366-300ffde2afb4" (UID: "b77a8d6f-5a83-494e-a366-300ffde2afb4"). InnerVolumeSpecName "kube-api-access-5pdnf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:52:04 crc kubenswrapper[4873]: I1201 08:52:04.687517 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b77a8d6f-5a83-494e-a366-300ffde2afb4-util" (OuterVolumeSpecName: "util") pod "b77a8d6f-5a83-494e-a366-300ffde2afb4" (UID: "b77a8d6f-5a83-494e-a366-300ffde2afb4"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:52:04 crc kubenswrapper[4873]: I1201 08:52:04.774466 4873 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b77a8d6f-5a83-494e-a366-300ffde2afb4-util\") on node \"crc\" DevicePath \"\"" Dec 01 08:52:04 crc kubenswrapper[4873]: I1201 08:52:04.774521 4873 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b77a8d6f-5a83-494e-a366-300ffde2afb4-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:52:04 crc kubenswrapper[4873]: I1201 08:52:04.774537 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5pdnf\" (UniqueName: \"kubernetes.io/projected/b77a8d6f-5a83-494e-a366-300ffde2afb4-kube-api-access-5pdnf\") on node \"crc\" DevicePath \"\"" Dec 01 08:52:05 crc kubenswrapper[4873]: I1201 08:52:05.345783 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhgjjb" event={"ID":"b77a8d6f-5a83-494e-a366-300ffde2afb4","Type":"ContainerDied","Data":"9129afb0477cab9f7df107dfa0be8546394ea628f920f598c1b30ffe113cf39c"} Dec 01 08:52:05 crc kubenswrapper[4873]: I1201 08:52:05.345824 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhgjjb" Dec 01 08:52:05 crc kubenswrapper[4873]: I1201 08:52:05.345854 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9129afb0477cab9f7df107dfa0be8546394ea628f920f598c1b30ffe113cf39c" Dec 01 08:52:06 crc kubenswrapper[4873]: I1201 08:52:06.956826 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-fbds9"] Dec 01 08:52:06 crc kubenswrapper[4873]: E1201 08:52:06.958943 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b77a8d6f-5a83-494e-a366-300ffde2afb4" containerName="util" Dec 01 08:52:06 crc kubenswrapper[4873]: I1201 08:52:06.959063 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="b77a8d6f-5a83-494e-a366-300ffde2afb4" containerName="util" Dec 01 08:52:06 crc kubenswrapper[4873]: E1201 08:52:06.959161 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b77a8d6f-5a83-494e-a366-300ffde2afb4" containerName="pull" Dec 01 08:52:06 crc kubenswrapper[4873]: I1201 08:52:06.959254 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="b77a8d6f-5a83-494e-a366-300ffde2afb4" containerName="pull" Dec 01 08:52:06 crc kubenswrapper[4873]: E1201 08:52:06.959351 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b77a8d6f-5a83-494e-a366-300ffde2afb4" containerName="extract" Dec 01 08:52:06 crc kubenswrapper[4873]: I1201 08:52:06.959432 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="b77a8d6f-5a83-494e-a366-300ffde2afb4" containerName="extract" Dec 01 08:52:06 crc kubenswrapper[4873]: I1201 08:52:06.959622 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="b77a8d6f-5a83-494e-a366-300ffde2afb4" containerName="extract" Dec 01 08:52:06 crc kubenswrapper[4873]: I1201 08:52:06.960252 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-fbds9" Dec 01 08:52:06 crc kubenswrapper[4873]: I1201 08:52:06.962914 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-mbzrd" Dec 01 08:52:06 crc kubenswrapper[4873]: I1201 08:52:06.963391 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 01 08:52:06 crc kubenswrapper[4873]: I1201 08:52:06.964659 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 01 08:52:06 crc kubenswrapper[4873]: I1201 08:52:06.967934 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-fbds9"] Dec 01 08:52:07 crc kubenswrapper[4873]: I1201 08:52:07.113243 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwrk9\" (UniqueName: \"kubernetes.io/projected/a61c7c72-a083-41c2-b0a4-707d26b095c6-kube-api-access-cwrk9\") pod \"nmstate-operator-5b5b58f5c8-fbds9\" (UID: \"a61c7c72-a083-41c2-b0a4-707d26b095c6\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-fbds9" Dec 01 08:52:07 crc kubenswrapper[4873]: I1201 08:52:07.214964 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwrk9\" (UniqueName: \"kubernetes.io/projected/a61c7c72-a083-41c2-b0a4-707d26b095c6-kube-api-access-cwrk9\") pod \"nmstate-operator-5b5b58f5c8-fbds9\" (UID: \"a61c7c72-a083-41c2-b0a4-707d26b095c6\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-fbds9" Dec 01 08:52:07 crc kubenswrapper[4873]: I1201 08:52:07.234994 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwrk9\" (UniqueName: \"kubernetes.io/projected/a61c7c72-a083-41c2-b0a4-707d26b095c6-kube-api-access-cwrk9\") pod \"nmstate-operator-5b5b58f5c8-fbds9\" (UID: \"a61c7c72-a083-41c2-b0a4-707d26b095c6\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-fbds9" Dec 01 08:52:07 crc kubenswrapper[4873]: I1201 08:52:07.277397 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-fbds9" Dec 01 08:52:07 crc kubenswrapper[4873]: I1201 08:52:07.489811 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-fbds9"] Dec 01 08:52:08 crc kubenswrapper[4873]: I1201 08:52:08.368926 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-fbds9" event={"ID":"a61c7c72-a083-41c2-b0a4-707d26b095c6","Type":"ContainerStarted","Data":"8ddd5e2f0b0eb19ce5aa2a72d775ee8dc6ac36e13901f9381e8db743e5fcc6d4"} Dec 01 08:52:10 crc kubenswrapper[4873]: I1201 08:52:10.386294 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-fbds9" event={"ID":"a61c7c72-a083-41c2-b0a4-707d26b095c6","Type":"ContainerStarted","Data":"8807a8cda71a623f246f4fb9612d28425ce066168ac87e7116b9843000149312"} Dec 01 08:52:10 crc kubenswrapper[4873]: I1201 08:52:10.404573 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-fbds9" podStartSLOduration=2.082624107 podStartE2EDuration="4.404542518s" podCreationTimestamp="2025-12-01 08:52:06 +0000 UTC" firstStartedPulling="2025-12-01 08:52:07.502080216 +0000 UTC m=+703.404188755" lastFinishedPulling="2025-12-01 08:52:09.823998627 +0000 UTC m=+705.726107166" observedRunningTime="2025-12-01 08:52:10.401775998 +0000 UTC m=+706.303884537" watchObservedRunningTime="2025-12-01 08:52:10.404542518 +0000 UTC m=+706.306651057" Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.385037 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-m7d79"] Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.386094 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-m7d79" Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.397947 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-m7d79"] Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.457035 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-78vtx" Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.458576 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-79fqx"] Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.459523 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-79fqx" Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.466027 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.479290 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-79fqx"] Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.489057 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/6050180f-b828-43a2-911c-4a44354d4bf6-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-79fqx\" (UID: \"6050180f-b828-43a2-911c-4a44354d4bf6\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-79fqx" Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.489133 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l458c\" (UniqueName: \"kubernetes.io/projected/6050180f-b828-43a2-911c-4a44354d4bf6-kube-api-access-l458c\") pod \"nmstate-webhook-5f6d4c5ccb-79fqx\" (UID: \"6050180f-b828-43a2-911c-4a44354d4bf6\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-79fqx" Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.489157 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnjbn\" (UniqueName: \"kubernetes.io/projected/1c32a501-0496-45b2-8e04-984bccb3c03d-kube-api-access-bnjbn\") pod \"nmstate-metrics-7f946cbc9-m7d79\" (UID: \"1c32a501-0496-45b2-8e04-984bccb3c03d\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-m7d79" Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.506666 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-x8slw"] Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.509746 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-x8slw" Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.590463 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/6050180f-b828-43a2-911c-4a44354d4bf6-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-79fqx\" (UID: \"6050180f-b828-43a2-911c-4a44354d4bf6\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-79fqx" Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.590546 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/5ca9fe36-a6a0-402c-a744-4125d0cd37c6-dbus-socket\") pod \"nmstate-handler-x8slw\" (UID: \"5ca9fe36-a6a0-402c-a744-4125d0cd37c6\") " pod="openshift-nmstate/nmstate-handler-x8slw" Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.590571 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l458c\" (UniqueName: \"kubernetes.io/projected/6050180f-b828-43a2-911c-4a44354d4bf6-kube-api-access-l458c\") pod \"nmstate-webhook-5f6d4c5ccb-79fqx\" (UID: \"6050180f-b828-43a2-911c-4a44354d4bf6\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-79fqx" Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.590598 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnjbn\" (UniqueName: \"kubernetes.io/projected/1c32a501-0496-45b2-8e04-984bccb3c03d-kube-api-access-bnjbn\") pod \"nmstate-metrics-7f946cbc9-m7d79\" (UID: \"1c32a501-0496-45b2-8e04-984bccb3c03d\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-m7d79" Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.590624 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/5ca9fe36-a6a0-402c-a744-4125d0cd37c6-nmstate-lock\") pod \"nmstate-handler-x8slw\" (UID: \"5ca9fe36-a6a0-402c-a744-4125d0cd37c6\") " pod="openshift-nmstate/nmstate-handler-x8slw" Dec 01 08:52:11 crc kubenswrapper[4873]: E1201 08:52:11.590635 4873 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Dec 01 08:52:11 crc kubenswrapper[4873]: E1201 08:52:11.590736 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6050180f-b828-43a2-911c-4a44354d4bf6-tls-key-pair podName:6050180f-b828-43a2-911c-4a44354d4bf6 nodeName:}" failed. No retries permitted until 2025-12-01 08:52:12.090709171 +0000 UTC m=+707.992817710 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/6050180f-b828-43a2-911c-4a44354d4bf6-tls-key-pair") pod "nmstate-webhook-5f6d4c5ccb-79fqx" (UID: "6050180f-b828-43a2-911c-4a44354d4bf6") : secret "openshift-nmstate-webhook" not found Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.590661 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/5ca9fe36-a6a0-402c-a744-4125d0cd37c6-ovs-socket\") pod \"nmstate-handler-x8slw\" (UID: \"5ca9fe36-a6a0-402c-a744-4125d0cd37c6\") " pod="openshift-nmstate/nmstate-handler-x8slw" Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.591069 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgdf8\" (UniqueName: \"kubernetes.io/projected/5ca9fe36-a6a0-402c-a744-4125d0cd37c6-kube-api-access-fgdf8\") pod \"nmstate-handler-x8slw\" (UID: \"5ca9fe36-a6a0-402c-a744-4125d0cd37c6\") " pod="openshift-nmstate/nmstate-handler-x8slw" Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.627363 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rtjft"] Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.628500 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rtjft" Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.630470 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnjbn\" (UniqueName: \"kubernetes.io/projected/1c32a501-0496-45b2-8e04-984bccb3c03d-kube-api-access-bnjbn\") pod \"nmstate-metrics-7f946cbc9-m7d79\" (UID: \"1c32a501-0496-45b2-8e04-984bccb3c03d\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-m7d79" Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.632441 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.632718 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.632765 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-dz8fg" Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.637261 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l458c\" (UniqueName: \"kubernetes.io/projected/6050180f-b828-43a2-911c-4a44354d4bf6-kube-api-access-l458c\") pod \"nmstate-webhook-5f6d4c5ccb-79fqx\" (UID: \"6050180f-b828-43a2-911c-4a44354d4bf6\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-79fqx" Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.637567 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rtjft"] Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.692382 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/5ca9fe36-a6a0-402c-a744-4125d0cd37c6-ovs-socket\") pod \"nmstate-handler-x8slw\" (UID: \"5ca9fe36-a6a0-402c-a744-4125d0cd37c6\") " pod="openshift-nmstate/nmstate-handler-x8slw" Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.692583 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/5ca9fe36-a6a0-402c-a744-4125d0cd37c6-ovs-socket\") pod \"nmstate-handler-x8slw\" (UID: \"5ca9fe36-a6a0-402c-a744-4125d0cd37c6\") " pod="openshift-nmstate/nmstate-handler-x8slw" Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.703405 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgdf8\" (UniqueName: \"kubernetes.io/projected/5ca9fe36-a6a0-402c-a744-4125d0cd37c6-kube-api-access-fgdf8\") pod \"nmstate-handler-x8slw\" (UID: \"5ca9fe36-a6a0-402c-a744-4125d0cd37c6\") " pod="openshift-nmstate/nmstate-handler-x8slw" Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.703487 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/244acac6-afd2-4d7b-b1a0-085a24cbf1c8-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-rtjft\" (UID: \"244acac6-afd2-4d7b-b1a0-085a24cbf1c8\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rtjft" Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.703573 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tk9d\" (UniqueName: \"kubernetes.io/projected/244acac6-afd2-4d7b-b1a0-085a24cbf1c8-kube-api-access-6tk9d\") pod \"nmstate-console-plugin-7fbb5f6569-rtjft\" (UID: \"244acac6-afd2-4d7b-b1a0-085a24cbf1c8\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rtjft" Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.703627 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/244acac6-afd2-4d7b-b1a0-085a24cbf1c8-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-rtjft\" (UID: \"244acac6-afd2-4d7b-b1a0-085a24cbf1c8\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rtjft" Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.703768 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/5ca9fe36-a6a0-402c-a744-4125d0cd37c6-dbus-socket\") pod \"nmstate-handler-x8slw\" (UID: \"5ca9fe36-a6a0-402c-a744-4125d0cd37c6\") " pod="openshift-nmstate/nmstate-handler-x8slw" Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.703851 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/5ca9fe36-a6a0-402c-a744-4125d0cd37c6-nmstate-lock\") pod \"nmstate-handler-x8slw\" (UID: \"5ca9fe36-a6a0-402c-a744-4125d0cd37c6\") " pod="openshift-nmstate/nmstate-handler-x8slw" Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.703947 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/5ca9fe36-a6a0-402c-a744-4125d0cd37c6-nmstate-lock\") pod \"nmstate-handler-x8slw\" (UID: \"5ca9fe36-a6a0-402c-a744-4125d0cd37c6\") " pod="openshift-nmstate/nmstate-handler-x8slw" Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.704327 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/5ca9fe36-a6a0-402c-a744-4125d0cd37c6-dbus-socket\") pod \"nmstate-handler-x8slw\" (UID: \"5ca9fe36-a6a0-402c-a744-4125d0cd37c6\") " pod="openshift-nmstate/nmstate-handler-x8slw" Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.725530 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgdf8\" (UniqueName: \"kubernetes.io/projected/5ca9fe36-a6a0-402c-a744-4125d0cd37c6-kube-api-access-fgdf8\") pod \"nmstate-handler-x8slw\" (UID: \"5ca9fe36-a6a0-402c-a744-4125d0cd37c6\") " pod="openshift-nmstate/nmstate-handler-x8slw" Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.774441 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-m7d79" Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.805422 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/244acac6-afd2-4d7b-b1a0-085a24cbf1c8-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-rtjft\" (UID: \"244acac6-afd2-4d7b-b1a0-085a24cbf1c8\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rtjft" Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.805580 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/244acac6-afd2-4d7b-b1a0-085a24cbf1c8-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-rtjft\" (UID: \"244acac6-afd2-4d7b-b1a0-085a24cbf1c8\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rtjft" Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.805628 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tk9d\" (UniqueName: \"kubernetes.io/projected/244acac6-afd2-4d7b-b1a0-085a24cbf1c8-kube-api-access-6tk9d\") pod \"nmstate-console-plugin-7fbb5f6569-rtjft\" (UID: \"244acac6-afd2-4d7b-b1a0-085a24cbf1c8\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rtjft" Dec 01 08:52:11 crc kubenswrapper[4873]: E1201 08:52:11.805788 4873 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Dec 01 08:52:11 crc kubenswrapper[4873]: E1201 08:52:11.805876 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/244acac6-afd2-4d7b-b1a0-085a24cbf1c8-plugin-serving-cert podName:244acac6-afd2-4d7b-b1a0-085a24cbf1c8 nodeName:}" failed. No retries permitted until 2025-12-01 08:52:12.305852207 +0000 UTC m=+708.207960746 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/244acac6-afd2-4d7b-b1a0-085a24cbf1c8-plugin-serving-cert") pod "nmstate-console-plugin-7fbb5f6569-rtjft" (UID: "244acac6-afd2-4d7b-b1a0-085a24cbf1c8") : secret "plugin-serving-cert" not found Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.806735 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/244acac6-afd2-4d7b-b1a0-085a24cbf1c8-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-rtjft\" (UID: \"244acac6-afd2-4d7b-b1a0-085a24cbf1c8\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rtjft" Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.818956 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-5c96596f67-54wd4"] Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.819870 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5c96596f67-54wd4" Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.829358 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tk9d\" (UniqueName: \"kubernetes.io/projected/244acac6-afd2-4d7b-b1a0-085a24cbf1c8-kube-api-access-6tk9d\") pod \"nmstate-console-plugin-7fbb5f6569-rtjft\" (UID: \"244acac6-afd2-4d7b-b1a0-085a24cbf1c8\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rtjft" Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.835828 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-x8slw" Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.844827 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5c96596f67-54wd4"] Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.906693 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cf865457-9e4e-49d3-8c4a-c74017ef82ff-trusted-ca-bundle\") pod \"console-5c96596f67-54wd4\" (UID: \"cf865457-9e4e-49d3-8c4a-c74017ef82ff\") " pod="openshift-console/console-5c96596f67-54wd4" Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.907173 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cf865457-9e4e-49d3-8c4a-c74017ef82ff-oauth-serving-cert\") pod \"console-5c96596f67-54wd4\" (UID: \"cf865457-9e4e-49d3-8c4a-c74017ef82ff\") " pod="openshift-console/console-5c96596f67-54wd4" Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.907198 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cf865457-9e4e-49d3-8c4a-c74017ef82ff-console-config\") pod \"console-5c96596f67-54wd4\" (UID: \"cf865457-9e4e-49d3-8c4a-c74017ef82ff\") " pod="openshift-console/console-5c96596f67-54wd4" Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.907228 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cf865457-9e4e-49d3-8c4a-c74017ef82ff-console-serving-cert\") pod \"console-5c96596f67-54wd4\" (UID: \"cf865457-9e4e-49d3-8c4a-c74017ef82ff\") " pod="openshift-console/console-5c96596f67-54wd4" Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.907262 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqgzt\" (UniqueName: \"kubernetes.io/projected/cf865457-9e4e-49d3-8c4a-c74017ef82ff-kube-api-access-qqgzt\") pod \"console-5c96596f67-54wd4\" (UID: \"cf865457-9e4e-49d3-8c4a-c74017ef82ff\") " pod="openshift-console/console-5c96596f67-54wd4" Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.907307 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cf865457-9e4e-49d3-8c4a-c74017ef82ff-service-ca\") pod \"console-5c96596f67-54wd4\" (UID: \"cf865457-9e4e-49d3-8c4a-c74017ef82ff\") " pod="openshift-console/console-5c96596f67-54wd4" Dec 01 08:52:11 crc kubenswrapper[4873]: I1201 08:52:11.907338 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cf865457-9e4e-49d3-8c4a-c74017ef82ff-console-oauth-config\") pod \"console-5c96596f67-54wd4\" (UID: \"cf865457-9e4e-49d3-8c4a-c74017ef82ff\") " pod="openshift-console/console-5c96596f67-54wd4" Dec 01 08:52:12 crc kubenswrapper[4873]: I1201 08:52:12.010441 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cf865457-9e4e-49d3-8c4a-c74017ef82ff-console-serving-cert\") pod \"console-5c96596f67-54wd4\" (UID: \"cf865457-9e4e-49d3-8c4a-c74017ef82ff\") " pod="openshift-console/console-5c96596f67-54wd4" Dec 01 08:52:12 crc kubenswrapper[4873]: I1201 08:52:12.010524 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqgzt\" (UniqueName: \"kubernetes.io/projected/cf865457-9e4e-49d3-8c4a-c74017ef82ff-kube-api-access-qqgzt\") pod \"console-5c96596f67-54wd4\" (UID: \"cf865457-9e4e-49d3-8c4a-c74017ef82ff\") " pod="openshift-console/console-5c96596f67-54wd4" Dec 01 08:52:12 crc kubenswrapper[4873]: I1201 08:52:12.010586 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cf865457-9e4e-49d3-8c4a-c74017ef82ff-service-ca\") pod \"console-5c96596f67-54wd4\" (UID: \"cf865457-9e4e-49d3-8c4a-c74017ef82ff\") " pod="openshift-console/console-5c96596f67-54wd4" Dec 01 08:52:12 crc kubenswrapper[4873]: I1201 08:52:12.010624 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cf865457-9e4e-49d3-8c4a-c74017ef82ff-console-oauth-config\") pod \"console-5c96596f67-54wd4\" (UID: \"cf865457-9e4e-49d3-8c4a-c74017ef82ff\") " pod="openshift-console/console-5c96596f67-54wd4" Dec 01 08:52:12 crc kubenswrapper[4873]: I1201 08:52:12.010654 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cf865457-9e4e-49d3-8c4a-c74017ef82ff-trusted-ca-bundle\") pod \"console-5c96596f67-54wd4\" (UID: \"cf865457-9e4e-49d3-8c4a-c74017ef82ff\") " pod="openshift-console/console-5c96596f67-54wd4" Dec 01 08:52:12 crc kubenswrapper[4873]: I1201 08:52:12.010696 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cf865457-9e4e-49d3-8c4a-c74017ef82ff-oauth-serving-cert\") pod \"console-5c96596f67-54wd4\" (UID: \"cf865457-9e4e-49d3-8c4a-c74017ef82ff\") " pod="openshift-console/console-5c96596f67-54wd4" Dec 01 08:52:12 crc kubenswrapper[4873]: I1201 08:52:12.010723 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cf865457-9e4e-49d3-8c4a-c74017ef82ff-console-config\") pod \"console-5c96596f67-54wd4\" (UID: \"cf865457-9e4e-49d3-8c4a-c74017ef82ff\") " pod="openshift-console/console-5c96596f67-54wd4" Dec 01 08:52:12 crc kubenswrapper[4873]: I1201 08:52:12.011986 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cf865457-9e4e-49d3-8c4a-c74017ef82ff-console-config\") pod \"console-5c96596f67-54wd4\" (UID: \"cf865457-9e4e-49d3-8c4a-c74017ef82ff\") " pod="openshift-console/console-5c96596f67-54wd4" Dec 01 08:52:12 crc kubenswrapper[4873]: I1201 08:52:12.012993 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cf865457-9e4e-49d3-8c4a-c74017ef82ff-service-ca\") pod \"console-5c96596f67-54wd4\" (UID: \"cf865457-9e4e-49d3-8c4a-c74017ef82ff\") " pod="openshift-console/console-5c96596f67-54wd4" Dec 01 08:52:12 crc kubenswrapper[4873]: I1201 08:52:12.013053 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cf865457-9e4e-49d3-8c4a-c74017ef82ff-oauth-serving-cert\") pod \"console-5c96596f67-54wd4\" (UID: \"cf865457-9e4e-49d3-8c4a-c74017ef82ff\") " pod="openshift-console/console-5c96596f67-54wd4" Dec 01 08:52:12 crc kubenswrapper[4873]: I1201 08:52:12.014341 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cf865457-9e4e-49d3-8c4a-c74017ef82ff-console-serving-cert\") pod \"console-5c96596f67-54wd4\" (UID: \"cf865457-9e4e-49d3-8c4a-c74017ef82ff\") " pod="openshift-console/console-5c96596f67-54wd4" Dec 01 08:52:12 crc kubenswrapper[4873]: I1201 08:52:12.014378 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cf865457-9e4e-49d3-8c4a-c74017ef82ff-trusted-ca-bundle\") pod \"console-5c96596f67-54wd4\" (UID: \"cf865457-9e4e-49d3-8c4a-c74017ef82ff\") " pod="openshift-console/console-5c96596f67-54wd4" Dec 01 08:52:12 crc kubenswrapper[4873]: I1201 08:52:12.015746 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cf865457-9e4e-49d3-8c4a-c74017ef82ff-console-oauth-config\") pod \"console-5c96596f67-54wd4\" (UID: \"cf865457-9e4e-49d3-8c4a-c74017ef82ff\") " pod="openshift-console/console-5c96596f67-54wd4" Dec 01 08:52:12 crc kubenswrapper[4873]: I1201 08:52:12.031692 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqgzt\" (UniqueName: \"kubernetes.io/projected/cf865457-9e4e-49d3-8c4a-c74017ef82ff-kube-api-access-qqgzt\") pod \"console-5c96596f67-54wd4\" (UID: \"cf865457-9e4e-49d3-8c4a-c74017ef82ff\") " pod="openshift-console/console-5c96596f67-54wd4" Dec 01 08:52:12 crc kubenswrapper[4873]: I1201 08:52:12.112394 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/6050180f-b828-43a2-911c-4a44354d4bf6-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-79fqx\" (UID: \"6050180f-b828-43a2-911c-4a44354d4bf6\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-79fqx" Dec 01 08:52:12 crc kubenswrapper[4873]: I1201 08:52:12.116169 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/6050180f-b828-43a2-911c-4a44354d4bf6-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-79fqx\" (UID: \"6050180f-b828-43a2-911c-4a44354d4bf6\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-79fqx" Dec 01 08:52:12 crc kubenswrapper[4873]: I1201 08:52:12.167197 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5c96596f67-54wd4" Dec 01 08:52:12 crc kubenswrapper[4873]: I1201 08:52:12.252827 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-m7d79"] Dec 01 08:52:12 crc kubenswrapper[4873]: W1201 08:52:12.265535 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1c32a501_0496_45b2_8e04_984bccb3c03d.slice/crio-faccf8a699d68fd015bcb78053af6a3ccf35998dea17c9f78e50644f829ee121 WatchSource:0}: Error finding container faccf8a699d68fd015bcb78053af6a3ccf35998dea17c9f78e50644f829ee121: Status 404 returned error can't find the container with id faccf8a699d68fd015bcb78053af6a3ccf35998dea17c9f78e50644f829ee121 Dec 01 08:52:12 crc kubenswrapper[4873]: I1201 08:52:12.314959 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/244acac6-afd2-4d7b-b1a0-085a24cbf1c8-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-rtjft\" (UID: \"244acac6-afd2-4d7b-b1a0-085a24cbf1c8\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rtjft" Dec 01 08:52:12 crc kubenswrapper[4873]: I1201 08:52:12.321870 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/244acac6-afd2-4d7b-b1a0-085a24cbf1c8-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-rtjft\" (UID: \"244acac6-afd2-4d7b-b1a0-085a24cbf1c8\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rtjft" Dec 01 08:52:12 crc kubenswrapper[4873]: I1201 08:52:12.368122 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5c96596f67-54wd4"] Dec 01 08:52:12 crc kubenswrapper[4873]: W1201 08:52:12.379428 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcf865457_9e4e_49d3_8c4a_c74017ef82ff.slice/crio-05378910b0b955ee2f8c4e08c8599ccb33050c615bbd4a28ca113fd8f617e21b WatchSource:0}: Error finding container 05378910b0b955ee2f8c4e08c8599ccb33050c615bbd4a28ca113fd8f617e21b: Status 404 returned error can't find the container with id 05378910b0b955ee2f8c4e08c8599ccb33050c615bbd4a28ca113fd8f617e21b Dec 01 08:52:12 crc kubenswrapper[4873]: I1201 08:52:12.392666 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-79fqx" Dec 01 08:52:12 crc kubenswrapper[4873]: I1201 08:52:12.400131 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-x8slw" event={"ID":"5ca9fe36-a6a0-402c-a744-4125d0cd37c6","Type":"ContainerStarted","Data":"5cb943d95c90faceba7533982ca2bece71752c27cb63818f967c1a87a140eeff"} Dec 01 08:52:12 crc kubenswrapper[4873]: I1201 08:52:12.402773 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-m7d79" event={"ID":"1c32a501-0496-45b2-8e04-984bccb3c03d","Type":"ContainerStarted","Data":"faccf8a699d68fd015bcb78053af6a3ccf35998dea17c9f78e50644f829ee121"} Dec 01 08:52:12 crc kubenswrapper[4873]: I1201 08:52:12.404598 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5c96596f67-54wd4" event={"ID":"cf865457-9e4e-49d3-8c4a-c74017ef82ff","Type":"ContainerStarted","Data":"05378910b0b955ee2f8c4e08c8599ccb33050c615bbd4a28ca113fd8f617e21b"} Dec 01 08:52:12 crc kubenswrapper[4873]: I1201 08:52:12.592309 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rtjft" Dec 01 08:52:12 crc kubenswrapper[4873]: I1201 08:52:12.805101 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rtjft"] Dec 01 08:52:12 crc kubenswrapper[4873]: I1201 08:52:12.879258 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-79fqx"] Dec 01 08:52:12 crc kubenswrapper[4873]: W1201 08:52:12.887102 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6050180f_b828_43a2_911c_4a44354d4bf6.slice/crio-803d2c5875e8255a0d44acd9fbfca079ae9ff917a511cc657f0c95f685f90ff2 WatchSource:0}: Error finding container 803d2c5875e8255a0d44acd9fbfca079ae9ff917a511cc657f0c95f685f90ff2: Status 404 returned error can't find the container with id 803d2c5875e8255a0d44acd9fbfca079ae9ff917a511cc657f0c95f685f90ff2 Dec 01 08:52:13 crc kubenswrapper[4873]: I1201 08:52:13.414874 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rtjft" event={"ID":"244acac6-afd2-4d7b-b1a0-085a24cbf1c8","Type":"ContainerStarted","Data":"a76dca3daaa834781e3bd1d3a1ec4eae5cfcba25982f8b3e753c9b4fe6f0016f"} Dec 01 08:52:13 crc kubenswrapper[4873]: I1201 08:52:13.418180 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-79fqx" event={"ID":"6050180f-b828-43a2-911c-4a44354d4bf6","Type":"ContainerStarted","Data":"803d2c5875e8255a0d44acd9fbfca079ae9ff917a511cc657f0c95f685f90ff2"} Dec 01 08:52:13 crc kubenswrapper[4873]: I1201 08:52:13.420733 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5c96596f67-54wd4" event={"ID":"cf865457-9e4e-49d3-8c4a-c74017ef82ff","Type":"ContainerStarted","Data":"10eb8e5a8d09ebdd9f60f947c14c2319a071da968f6456a76d0256a88b6df32a"} Dec 01 08:52:13 crc kubenswrapper[4873]: I1201 08:52:13.449981 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-5c96596f67-54wd4" podStartSLOduration=2.4499489 podStartE2EDuration="2.4499489s" podCreationTimestamp="2025-12-01 08:52:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:52:13.442275155 +0000 UTC m=+709.344383694" watchObservedRunningTime="2025-12-01 08:52:13.4499489 +0000 UTC m=+709.352057439" Dec 01 08:52:15 crc kubenswrapper[4873]: I1201 08:52:15.438292 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-x8slw" event={"ID":"5ca9fe36-a6a0-402c-a744-4125d0cd37c6","Type":"ContainerStarted","Data":"cf0c61da563564af058008a978ba776e83b0458410af6bf64c6fda43ce39a730"} Dec 01 08:52:15 crc kubenswrapper[4873]: I1201 08:52:15.439123 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-x8slw" Dec 01 08:52:15 crc kubenswrapper[4873]: I1201 08:52:15.441466 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-m7d79" event={"ID":"1c32a501-0496-45b2-8e04-984bccb3c03d","Type":"ContainerStarted","Data":"9d4a0a44959042c0cf3a64218105ec74d1077ea1c71b6f1dfdf6ab6ed57a8117"} Dec 01 08:52:15 crc kubenswrapper[4873]: I1201 08:52:15.447549 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-79fqx" event={"ID":"6050180f-b828-43a2-911c-4a44354d4bf6","Type":"ContainerStarted","Data":"1f426cb47a57a93ad015c44741fee715934181ec3c4ce5b89fdebcb77067c82b"} Dec 01 08:52:15 crc kubenswrapper[4873]: I1201 08:52:15.448670 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-79fqx" Dec 01 08:52:15 crc kubenswrapper[4873]: I1201 08:52:15.483248 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-x8slw" podStartSLOduration=1.8218333530000002 podStartE2EDuration="4.483213249s" podCreationTimestamp="2025-12-01 08:52:11 +0000 UTC" firstStartedPulling="2025-12-01 08:52:11.87837514 +0000 UTC m=+707.780483679" lastFinishedPulling="2025-12-01 08:52:14.539755036 +0000 UTC m=+710.441863575" observedRunningTime="2025-12-01 08:52:15.458977477 +0000 UTC m=+711.361086016" watchObservedRunningTime="2025-12-01 08:52:15.483213249 +0000 UTC m=+711.385321818" Dec 01 08:52:15 crc kubenswrapper[4873]: I1201 08:52:15.484599 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-79fqx" podStartSLOduration=2.824528083 podStartE2EDuration="4.484589859s" podCreationTimestamp="2025-12-01 08:52:11 +0000 UTC" firstStartedPulling="2025-12-01 08:52:12.889398849 +0000 UTC m=+708.791507388" lastFinishedPulling="2025-12-01 08:52:14.549460625 +0000 UTC m=+710.451569164" observedRunningTime="2025-12-01 08:52:15.476982255 +0000 UTC m=+711.379090794" watchObservedRunningTime="2025-12-01 08:52:15.484589859 +0000 UTC m=+711.386698428" Dec 01 08:52:16 crc kubenswrapper[4873]: I1201 08:52:16.461274 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rtjft" event={"ID":"244acac6-afd2-4d7b-b1a0-085a24cbf1c8","Type":"ContainerStarted","Data":"f56030e664295587cba917eab35044d85092c2e0a13eb9fd71d45f90d98b024a"} Dec 01 08:52:16 crc kubenswrapper[4873]: I1201 08:52:16.488724 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rtjft" podStartSLOduration=2.693829557 podStartE2EDuration="5.488693019s" podCreationTimestamp="2025-12-01 08:52:11 +0000 UTC" firstStartedPulling="2025-12-01 08:52:12.817581572 +0000 UTC m=+708.719690111" lastFinishedPulling="2025-12-01 08:52:15.612445034 +0000 UTC m=+711.514553573" observedRunningTime="2025-12-01 08:52:16.482972085 +0000 UTC m=+712.385080664" watchObservedRunningTime="2025-12-01 08:52:16.488693019 +0000 UTC m=+712.390801558" Dec 01 08:52:18 crc kubenswrapper[4873]: I1201 08:52:18.475155 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-m7d79" event={"ID":"1c32a501-0496-45b2-8e04-984bccb3c03d","Type":"ContainerStarted","Data":"acf2a136fa0bc7b1ab31ee119d0b470e04b373deff9d90ebbf68eaddd48388ec"} Dec 01 08:52:18 crc kubenswrapper[4873]: I1201 08:52:18.493279 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-m7d79" podStartSLOduration=2.25787169 podStartE2EDuration="7.493251849s" podCreationTimestamp="2025-12-01 08:52:11 +0000 UTC" firstStartedPulling="2025-12-01 08:52:12.267452096 +0000 UTC m=+708.169560635" lastFinishedPulling="2025-12-01 08:52:17.502832235 +0000 UTC m=+713.404940794" observedRunningTime="2025-12-01 08:52:18.493033844 +0000 UTC m=+714.395142403" watchObservedRunningTime="2025-12-01 08:52:18.493251849 +0000 UTC m=+714.395360398" Dec 01 08:52:21 crc kubenswrapper[4873]: I1201 08:52:21.859834 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-x8slw" Dec 01 08:52:22 crc kubenswrapper[4873]: I1201 08:52:22.167894 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-5c96596f67-54wd4" Dec 01 08:52:22 crc kubenswrapper[4873]: I1201 08:52:22.168327 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-5c96596f67-54wd4" Dec 01 08:52:22 crc kubenswrapper[4873]: I1201 08:52:22.174201 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-5c96596f67-54wd4" Dec 01 08:52:22 crc kubenswrapper[4873]: I1201 08:52:22.526554 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-5c96596f67-54wd4" Dec 01 08:52:22 crc kubenswrapper[4873]: I1201 08:52:22.580708 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-292nl"] Dec 01 08:52:31 crc kubenswrapper[4873]: I1201 08:52:31.059216 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:52:31 crc kubenswrapper[4873]: I1201 08:52:31.059769 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:52:32 crc kubenswrapper[4873]: I1201 08:52:32.399084 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-79fqx" Dec 01 08:52:47 crc kubenswrapper[4873]: I1201 08:52:47.627542 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-292nl" podUID="9491cb1f-acf6-438b-8175-11737d7bd245" containerName="console" containerID="cri-o://b078318cd9203fe5f24f2bd60ba9cea7d9e49405241d9f9008a2b5e2924098f8" gracePeriod=15 Dec 01 08:52:47 crc kubenswrapper[4873]: I1201 08:52:47.974519 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8359xjq"] Dec 01 08:52:47 crc kubenswrapper[4873]: I1201 08:52:47.977308 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8359xjq" Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:47.990658 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:48.012255 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8359xjq"] Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:48.114273 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-292nl_9491cb1f-acf6-438b-8175-11737d7bd245/console/0.log" Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:48.114365 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-292nl" Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:48.114511 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8d9127f2-7870-4647-8655-0c8fff4df500-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8359xjq\" (UID: \"8d9127f2-7870-4647-8655-0c8fff4df500\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8359xjq" Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:48.114605 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8d9127f2-7870-4647-8655-0c8fff4df500-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8359xjq\" (UID: \"8d9127f2-7870-4647-8655-0c8fff4df500\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8359xjq" Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:48.114674 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7q7fs\" (UniqueName: \"kubernetes.io/projected/8d9127f2-7870-4647-8655-0c8fff4df500-kube-api-access-7q7fs\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8359xjq\" (UID: \"8d9127f2-7870-4647-8655-0c8fff4df500\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8359xjq" Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:48.215833 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9491cb1f-acf6-438b-8175-11737d7bd245-oauth-serving-cert\") pod \"9491cb1f-acf6-438b-8175-11737d7bd245\" (UID: \"9491cb1f-acf6-438b-8175-11737d7bd245\") " Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:48.215953 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9491cb1f-acf6-438b-8175-11737d7bd245-console-config\") pod \"9491cb1f-acf6-438b-8175-11737d7bd245\" (UID: \"9491cb1f-acf6-438b-8175-11737d7bd245\") " Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:48.216003 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9491cb1f-acf6-438b-8175-11737d7bd245-service-ca\") pod \"9491cb1f-acf6-438b-8175-11737d7bd245\" (UID: \"9491cb1f-acf6-438b-8175-11737d7bd245\") " Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:48.216057 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9491cb1f-acf6-438b-8175-11737d7bd245-trusted-ca-bundle\") pod \"9491cb1f-acf6-438b-8175-11737d7bd245\" (UID: \"9491cb1f-acf6-438b-8175-11737d7bd245\") " Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:48.216106 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wggrb\" (UniqueName: \"kubernetes.io/projected/9491cb1f-acf6-438b-8175-11737d7bd245-kube-api-access-wggrb\") pod \"9491cb1f-acf6-438b-8175-11737d7bd245\" (UID: \"9491cb1f-acf6-438b-8175-11737d7bd245\") " Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:48.216153 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9491cb1f-acf6-438b-8175-11737d7bd245-console-serving-cert\") pod \"9491cb1f-acf6-438b-8175-11737d7bd245\" (UID: \"9491cb1f-acf6-438b-8175-11737d7bd245\") " Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:48.216189 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9491cb1f-acf6-438b-8175-11737d7bd245-console-oauth-config\") pod \"9491cb1f-acf6-438b-8175-11737d7bd245\" (UID: \"9491cb1f-acf6-438b-8175-11737d7bd245\") " Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:48.216358 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8d9127f2-7870-4647-8655-0c8fff4df500-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8359xjq\" (UID: \"8d9127f2-7870-4647-8655-0c8fff4df500\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8359xjq" Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:48.216403 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8d9127f2-7870-4647-8655-0c8fff4df500-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8359xjq\" (UID: \"8d9127f2-7870-4647-8655-0c8fff4df500\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8359xjq" Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:48.216482 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7q7fs\" (UniqueName: \"kubernetes.io/projected/8d9127f2-7870-4647-8655-0c8fff4df500-kube-api-access-7q7fs\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8359xjq\" (UID: \"8d9127f2-7870-4647-8655-0c8fff4df500\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8359xjq" Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:48.216986 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9491cb1f-acf6-438b-8175-11737d7bd245-console-config" (OuterVolumeSpecName: "console-config") pod "9491cb1f-acf6-438b-8175-11737d7bd245" (UID: "9491cb1f-acf6-438b-8175-11737d7bd245"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:48.216997 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9491cb1f-acf6-438b-8175-11737d7bd245-service-ca" (OuterVolumeSpecName: "service-ca") pod "9491cb1f-acf6-438b-8175-11737d7bd245" (UID: "9491cb1f-acf6-438b-8175-11737d7bd245"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:48.217530 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8d9127f2-7870-4647-8655-0c8fff4df500-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8359xjq\" (UID: \"8d9127f2-7870-4647-8655-0c8fff4df500\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8359xjq" Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:48.217664 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9491cb1f-acf6-438b-8175-11737d7bd245-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "9491cb1f-acf6-438b-8175-11737d7bd245" (UID: "9491cb1f-acf6-438b-8175-11737d7bd245"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:48.217787 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8d9127f2-7870-4647-8655-0c8fff4df500-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8359xjq\" (UID: \"8d9127f2-7870-4647-8655-0c8fff4df500\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8359xjq" Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:48.217915 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9491cb1f-acf6-438b-8175-11737d7bd245-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "9491cb1f-acf6-438b-8175-11737d7bd245" (UID: "9491cb1f-acf6-438b-8175-11737d7bd245"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:48.224516 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9491cb1f-acf6-438b-8175-11737d7bd245-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "9491cb1f-acf6-438b-8175-11737d7bd245" (UID: "9491cb1f-acf6-438b-8175-11737d7bd245"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:48.224889 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9491cb1f-acf6-438b-8175-11737d7bd245-kube-api-access-wggrb" (OuterVolumeSpecName: "kube-api-access-wggrb") pod "9491cb1f-acf6-438b-8175-11737d7bd245" (UID: "9491cb1f-acf6-438b-8175-11737d7bd245"). InnerVolumeSpecName "kube-api-access-wggrb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:48.225262 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9491cb1f-acf6-438b-8175-11737d7bd245-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "9491cb1f-acf6-438b-8175-11737d7bd245" (UID: "9491cb1f-acf6-438b-8175-11737d7bd245"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:48.241131 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7q7fs\" (UniqueName: \"kubernetes.io/projected/8d9127f2-7870-4647-8655-0c8fff4df500-kube-api-access-7q7fs\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8359xjq\" (UID: \"8d9127f2-7870-4647-8655-0c8fff4df500\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8359xjq" Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:48.318251 4873 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9491cb1f-acf6-438b-8175-11737d7bd245-console-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:48.318311 4873 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9491cb1f-acf6-438b-8175-11737d7bd245-service-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:48.318326 4873 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9491cb1f-acf6-438b-8175-11737d7bd245-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:48.318340 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wggrb\" (UniqueName: \"kubernetes.io/projected/9491cb1f-acf6-438b-8175-11737d7bd245-kube-api-access-wggrb\") on node \"crc\" DevicePath \"\"" Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:48.318364 4873 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9491cb1f-acf6-438b-8175-11737d7bd245-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:48.318376 4873 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9491cb1f-acf6-438b-8175-11737d7bd245-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:48.318390 4873 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9491cb1f-acf6-438b-8175-11737d7bd245-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:48.329601 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8359xjq" Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:48.546814 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8359xjq"] Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:48.688581 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8359xjq" event={"ID":"8d9127f2-7870-4647-8655-0c8fff4df500","Type":"ContainerStarted","Data":"95caae399c4ddba6f8707d96f40f52d45bddb51fc1f510a176c48a498ef572ac"} Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:48.689131 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8359xjq" event={"ID":"8d9127f2-7870-4647-8655-0c8fff4df500","Type":"ContainerStarted","Data":"5d96ce28b201ba3ba597e2b6259f0c86ecd378ab089ba8f5e4c934d32e8d96cc"} Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:48.690807 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-292nl_9491cb1f-acf6-438b-8175-11737d7bd245/console/0.log" Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:48.690849 4873 generic.go:334] "Generic (PLEG): container finished" podID="9491cb1f-acf6-438b-8175-11737d7bd245" containerID="b078318cd9203fe5f24f2bd60ba9cea7d9e49405241d9f9008a2b5e2924098f8" exitCode=2 Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:48.690876 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-292nl" event={"ID":"9491cb1f-acf6-438b-8175-11737d7bd245","Type":"ContainerDied","Data":"b078318cd9203fe5f24f2bd60ba9cea7d9e49405241d9f9008a2b5e2924098f8"} Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:48.690899 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-292nl" event={"ID":"9491cb1f-acf6-438b-8175-11737d7bd245","Type":"ContainerDied","Data":"297abfad07bfd6ea781b5c866c64d514efa3d17ffa8ee136a685828d797de4a3"} Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:48.690911 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-292nl" Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:48.690917 4873 scope.go:117] "RemoveContainer" containerID="b078318cd9203fe5f24f2bd60ba9cea7d9e49405241d9f9008a2b5e2924098f8" Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:48.714071 4873 scope.go:117] "RemoveContainer" containerID="b078318cd9203fe5f24f2bd60ba9cea7d9e49405241d9f9008a2b5e2924098f8" Dec 01 08:52:48 crc kubenswrapper[4873]: E1201 08:52:48.714658 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b078318cd9203fe5f24f2bd60ba9cea7d9e49405241d9f9008a2b5e2924098f8\": container with ID starting with b078318cd9203fe5f24f2bd60ba9cea7d9e49405241d9f9008a2b5e2924098f8 not found: ID does not exist" containerID="b078318cd9203fe5f24f2bd60ba9cea7d9e49405241d9f9008a2b5e2924098f8" Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:48.714730 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b078318cd9203fe5f24f2bd60ba9cea7d9e49405241d9f9008a2b5e2924098f8"} err="failed to get container status \"b078318cd9203fe5f24f2bd60ba9cea7d9e49405241d9f9008a2b5e2924098f8\": rpc error: code = NotFound desc = could not find container \"b078318cd9203fe5f24f2bd60ba9cea7d9e49405241d9f9008a2b5e2924098f8\": container with ID starting with b078318cd9203fe5f24f2bd60ba9cea7d9e49405241d9f9008a2b5e2924098f8 not found: ID does not exist" Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:48.727551 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-292nl"] Dec 01 08:52:48 crc kubenswrapper[4873]: I1201 08:52:48.731093 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-292nl"] Dec 01 08:52:49 crc kubenswrapper[4873]: I1201 08:52:49.707602 4873 generic.go:334] "Generic (PLEG): container finished" podID="8d9127f2-7870-4647-8655-0c8fff4df500" containerID="95caae399c4ddba6f8707d96f40f52d45bddb51fc1f510a176c48a498ef572ac" exitCode=0 Dec 01 08:52:49 crc kubenswrapper[4873]: I1201 08:52:49.707731 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8359xjq" event={"ID":"8d9127f2-7870-4647-8655-0c8fff4df500","Type":"ContainerDied","Data":"95caae399c4ddba6f8707d96f40f52d45bddb51fc1f510a176c48a498ef572ac"} Dec 01 08:52:50 crc kubenswrapper[4873]: I1201 08:52:50.439583 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9491cb1f-acf6-438b-8175-11737d7bd245" path="/var/lib/kubelet/pods/9491cb1f-acf6-438b-8175-11737d7bd245/volumes" Dec 01 08:52:51 crc kubenswrapper[4873]: I1201 08:52:51.723984 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8359xjq" event={"ID":"8d9127f2-7870-4647-8655-0c8fff4df500","Type":"ContainerStarted","Data":"b9ae590d382469552c5750909067e1e7c1e3099fba355311e2d230860e41b380"} Dec 01 08:52:52 crc kubenswrapper[4873]: I1201 08:52:52.732950 4873 generic.go:334] "Generic (PLEG): container finished" podID="8d9127f2-7870-4647-8655-0c8fff4df500" containerID="b9ae590d382469552c5750909067e1e7c1e3099fba355311e2d230860e41b380" exitCode=0 Dec 01 08:52:52 crc kubenswrapper[4873]: I1201 08:52:52.733124 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8359xjq" event={"ID":"8d9127f2-7870-4647-8655-0c8fff4df500","Type":"ContainerDied","Data":"b9ae590d382469552c5750909067e1e7c1e3099fba355311e2d230860e41b380"} Dec 01 08:52:53 crc kubenswrapper[4873]: I1201 08:52:53.744363 4873 generic.go:334] "Generic (PLEG): container finished" podID="8d9127f2-7870-4647-8655-0c8fff4df500" containerID="af307e44cebac494468b61c3f20b55706604086644e3cc574c48bb0d6a8221f5" exitCode=0 Dec 01 08:52:53 crc kubenswrapper[4873]: I1201 08:52:53.744444 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8359xjq" event={"ID":"8d9127f2-7870-4647-8655-0c8fff4df500","Type":"ContainerDied","Data":"af307e44cebac494468b61c3f20b55706604086644e3cc574c48bb0d6a8221f5"} Dec 01 08:52:54 crc kubenswrapper[4873]: I1201 08:52:54.996257 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8359xjq" Dec 01 08:52:55 crc kubenswrapper[4873]: I1201 08:52:55.122852 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8d9127f2-7870-4647-8655-0c8fff4df500-bundle\") pod \"8d9127f2-7870-4647-8655-0c8fff4df500\" (UID: \"8d9127f2-7870-4647-8655-0c8fff4df500\") " Dec 01 08:52:55 crc kubenswrapper[4873]: I1201 08:52:55.123054 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8d9127f2-7870-4647-8655-0c8fff4df500-util\") pod \"8d9127f2-7870-4647-8655-0c8fff4df500\" (UID: \"8d9127f2-7870-4647-8655-0c8fff4df500\") " Dec 01 08:52:55 crc kubenswrapper[4873]: I1201 08:52:55.123117 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7q7fs\" (UniqueName: \"kubernetes.io/projected/8d9127f2-7870-4647-8655-0c8fff4df500-kube-api-access-7q7fs\") pod \"8d9127f2-7870-4647-8655-0c8fff4df500\" (UID: \"8d9127f2-7870-4647-8655-0c8fff4df500\") " Dec 01 08:52:55 crc kubenswrapper[4873]: I1201 08:52:55.124622 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d9127f2-7870-4647-8655-0c8fff4df500-bundle" (OuterVolumeSpecName: "bundle") pod "8d9127f2-7870-4647-8655-0c8fff4df500" (UID: "8d9127f2-7870-4647-8655-0c8fff4df500"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:52:55 crc kubenswrapper[4873]: I1201 08:52:55.132781 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d9127f2-7870-4647-8655-0c8fff4df500-kube-api-access-7q7fs" (OuterVolumeSpecName: "kube-api-access-7q7fs") pod "8d9127f2-7870-4647-8655-0c8fff4df500" (UID: "8d9127f2-7870-4647-8655-0c8fff4df500"). InnerVolumeSpecName "kube-api-access-7q7fs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:52:55 crc kubenswrapper[4873]: I1201 08:52:55.149154 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d9127f2-7870-4647-8655-0c8fff4df500-util" (OuterVolumeSpecName: "util") pod "8d9127f2-7870-4647-8655-0c8fff4df500" (UID: "8d9127f2-7870-4647-8655-0c8fff4df500"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:52:55 crc kubenswrapper[4873]: I1201 08:52:55.224985 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7q7fs\" (UniqueName: \"kubernetes.io/projected/8d9127f2-7870-4647-8655-0c8fff4df500-kube-api-access-7q7fs\") on node \"crc\" DevicePath \"\"" Dec 01 08:52:55 crc kubenswrapper[4873]: I1201 08:52:55.225206 4873 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8d9127f2-7870-4647-8655-0c8fff4df500-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:52:55 crc kubenswrapper[4873]: I1201 08:52:55.225233 4873 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8d9127f2-7870-4647-8655-0c8fff4df500-util\") on node \"crc\" DevicePath \"\"" Dec 01 08:52:55 crc kubenswrapper[4873]: I1201 08:52:55.766913 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8359xjq" event={"ID":"8d9127f2-7870-4647-8655-0c8fff4df500","Type":"ContainerDied","Data":"5d96ce28b201ba3ba597e2b6259f0c86ecd378ab089ba8f5e4c934d32e8d96cc"} Dec 01 08:52:55 crc kubenswrapper[4873]: I1201 08:52:55.767339 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d96ce28b201ba3ba597e2b6259f0c86ecd378ab089ba8f5e4c934d32e8d96cc" Dec 01 08:52:55 crc kubenswrapper[4873]: I1201 08:52:55.766995 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8359xjq" Dec 01 08:53:00 crc kubenswrapper[4873]: I1201 08:53:00.374356 4873 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 01 08:53:01 crc kubenswrapper[4873]: I1201 08:53:01.059546 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:53:01 crc kubenswrapper[4873]: I1201 08:53:01.059625 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:53:01 crc kubenswrapper[4873]: I1201 08:53:01.059690 4873 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" Dec 01 08:53:01 crc kubenswrapper[4873]: I1201 08:53:01.060548 4873 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6650a48ef92f241e9c56ee0be1ecde624a461ff74d058cf63149c0cd17ae8d2d"} pod="openshift-machine-config-operator/machine-config-daemon-scwpp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 08:53:01 crc kubenswrapper[4873]: I1201 08:53:01.060624 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" containerID="cri-o://6650a48ef92f241e9c56ee0be1ecde624a461ff74d058cf63149c0cd17ae8d2d" gracePeriod=600 Dec 01 08:53:01 crc kubenswrapper[4873]: I1201 08:53:01.809280 4873 generic.go:334] "Generic (PLEG): container finished" podID="fef7b114-0e07-402d-a37b-315c36011f4b" containerID="6650a48ef92f241e9c56ee0be1ecde624a461ff74d058cf63149c0cd17ae8d2d" exitCode=0 Dec 01 08:53:01 crc kubenswrapper[4873]: I1201 08:53:01.809370 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" event={"ID":"fef7b114-0e07-402d-a37b-315c36011f4b","Type":"ContainerDied","Data":"6650a48ef92f241e9c56ee0be1ecde624a461ff74d058cf63149c0cd17ae8d2d"} Dec 01 08:53:01 crc kubenswrapper[4873]: I1201 08:53:01.809953 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" event={"ID":"fef7b114-0e07-402d-a37b-315c36011f4b","Type":"ContainerStarted","Data":"2af36981211968ab2b3e6fc266a51d247e83e9c0c856f6566cda62f7eeb85b9a"} Dec 01 08:53:01 crc kubenswrapper[4873]: I1201 08:53:01.810000 4873 scope.go:117] "RemoveContainer" containerID="0d13b8cddab09d2b96e4bfebd3f2cc00fd3cfd06d73474412b9d0387a8e1d7e0" Dec 01 08:53:06 crc kubenswrapper[4873]: I1201 08:53:06.071491 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-586dc797b5-msvs2"] Dec 01 08:53:06 crc kubenswrapper[4873]: E1201 08:53:06.072106 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9491cb1f-acf6-438b-8175-11737d7bd245" containerName="console" Dec 01 08:53:06 crc kubenswrapper[4873]: I1201 08:53:06.072120 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="9491cb1f-acf6-438b-8175-11737d7bd245" containerName="console" Dec 01 08:53:06 crc kubenswrapper[4873]: E1201 08:53:06.072144 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d9127f2-7870-4647-8655-0c8fff4df500" containerName="util" Dec 01 08:53:06 crc kubenswrapper[4873]: I1201 08:53:06.072149 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d9127f2-7870-4647-8655-0c8fff4df500" containerName="util" Dec 01 08:53:06 crc kubenswrapper[4873]: E1201 08:53:06.072159 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d9127f2-7870-4647-8655-0c8fff4df500" containerName="pull" Dec 01 08:53:06 crc kubenswrapper[4873]: I1201 08:53:06.072166 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d9127f2-7870-4647-8655-0c8fff4df500" containerName="pull" Dec 01 08:53:06 crc kubenswrapper[4873]: E1201 08:53:06.072176 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d9127f2-7870-4647-8655-0c8fff4df500" containerName="extract" Dec 01 08:53:06 crc kubenswrapper[4873]: I1201 08:53:06.072181 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d9127f2-7870-4647-8655-0c8fff4df500" containerName="extract" Dec 01 08:53:06 crc kubenswrapper[4873]: I1201 08:53:06.072283 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="9491cb1f-acf6-438b-8175-11737d7bd245" containerName="console" Dec 01 08:53:06 crc kubenswrapper[4873]: I1201 08:53:06.072292 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d9127f2-7870-4647-8655-0c8fff4df500" containerName="extract" Dec 01 08:53:06 crc kubenswrapper[4873]: I1201 08:53:06.072788 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-586dc797b5-msvs2" Dec 01 08:53:06 crc kubenswrapper[4873]: I1201 08:53:06.081887 4873 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 01 08:53:06 crc kubenswrapper[4873]: I1201 08:53:06.082137 4873 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 01 08:53:06 crc kubenswrapper[4873]: I1201 08:53:06.083266 4873 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-x5d2j" Dec 01 08:53:06 crc kubenswrapper[4873]: I1201 08:53:06.090117 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 01 08:53:06 crc kubenswrapper[4873]: I1201 08:53:06.090442 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 01 08:53:06 crc kubenswrapper[4873]: I1201 08:53:06.106774 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-586dc797b5-msvs2"] Dec 01 08:53:06 crc kubenswrapper[4873]: I1201 08:53:06.189989 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5b1a0fb8-99da-4202-8c8e-6f5fbcec1908-webhook-cert\") pod \"metallb-operator-controller-manager-586dc797b5-msvs2\" (UID: \"5b1a0fb8-99da-4202-8c8e-6f5fbcec1908\") " pod="metallb-system/metallb-operator-controller-manager-586dc797b5-msvs2" Dec 01 08:53:06 crc kubenswrapper[4873]: I1201 08:53:06.190399 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5b1a0fb8-99da-4202-8c8e-6f5fbcec1908-apiservice-cert\") pod \"metallb-operator-controller-manager-586dc797b5-msvs2\" (UID: \"5b1a0fb8-99da-4202-8c8e-6f5fbcec1908\") " pod="metallb-system/metallb-operator-controller-manager-586dc797b5-msvs2" Dec 01 08:53:06 crc kubenswrapper[4873]: I1201 08:53:06.190435 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57j7z\" (UniqueName: \"kubernetes.io/projected/5b1a0fb8-99da-4202-8c8e-6f5fbcec1908-kube-api-access-57j7z\") pod \"metallb-operator-controller-manager-586dc797b5-msvs2\" (UID: \"5b1a0fb8-99da-4202-8c8e-6f5fbcec1908\") " pod="metallb-system/metallb-operator-controller-manager-586dc797b5-msvs2" Dec 01 08:53:06 crc kubenswrapper[4873]: I1201 08:53:06.292043 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5b1a0fb8-99da-4202-8c8e-6f5fbcec1908-webhook-cert\") pod \"metallb-operator-controller-manager-586dc797b5-msvs2\" (UID: \"5b1a0fb8-99da-4202-8c8e-6f5fbcec1908\") " pod="metallb-system/metallb-operator-controller-manager-586dc797b5-msvs2" Dec 01 08:53:06 crc kubenswrapper[4873]: I1201 08:53:06.292124 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5b1a0fb8-99da-4202-8c8e-6f5fbcec1908-apiservice-cert\") pod \"metallb-operator-controller-manager-586dc797b5-msvs2\" (UID: \"5b1a0fb8-99da-4202-8c8e-6f5fbcec1908\") " pod="metallb-system/metallb-operator-controller-manager-586dc797b5-msvs2" Dec 01 08:53:06 crc kubenswrapper[4873]: I1201 08:53:06.292163 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57j7z\" (UniqueName: \"kubernetes.io/projected/5b1a0fb8-99da-4202-8c8e-6f5fbcec1908-kube-api-access-57j7z\") pod \"metallb-operator-controller-manager-586dc797b5-msvs2\" (UID: \"5b1a0fb8-99da-4202-8c8e-6f5fbcec1908\") " pod="metallb-system/metallb-operator-controller-manager-586dc797b5-msvs2" Dec 01 08:53:06 crc kubenswrapper[4873]: I1201 08:53:06.299166 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5b1a0fb8-99da-4202-8c8e-6f5fbcec1908-webhook-cert\") pod \"metallb-operator-controller-manager-586dc797b5-msvs2\" (UID: \"5b1a0fb8-99da-4202-8c8e-6f5fbcec1908\") " pod="metallb-system/metallb-operator-controller-manager-586dc797b5-msvs2" Dec 01 08:53:06 crc kubenswrapper[4873]: I1201 08:53:06.299192 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5b1a0fb8-99da-4202-8c8e-6f5fbcec1908-apiservice-cert\") pod \"metallb-operator-controller-manager-586dc797b5-msvs2\" (UID: \"5b1a0fb8-99da-4202-8c8e-6f5fbcec1908\") " pod="metallb-system/metallb-operator-controller-manager-586dc797b5-msvs2" Dec 01 08:53:06 crc kubenswrapper[4873]: I1201 08:53:06.320895 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57j7z\" (UniqueName: \"kubernetes.io/projected/5b1a0fb8-99da-4202-8c8e-6f5fbcec1908-kube-api-access-57j7z\") pod \"metallb-operator-controller-manager-586dc797b5-msvs2\" (UID: \"5b1a0fb8-99da-4202-8c8e-6f5fbcec1908\") " pod="metallb-system/metallb-operator-controller-manager-586dc797b5-msvs2" Dec 01 08:53:06 crc kubenswrapper[4873]: I1201 08:53:06.351314 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-767c58684d-z249t"] Dec 01 08:53:06 crc kubenswrapper[4873]: I1201 08:53:06.352382 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-767c58684d-z249t" Dec 01 08:53:06 crc kubenswrapper[4873]: I1201 08:53:06.354959 4873 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-z62mh" Dec 01 08:53:06 crc kubenswrapper[4873]: I1201 08:53:06.355072 4873 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 01 08:53:06 crc kubenswrapper[4873]: I1201 08:53:06.355120 4873 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 01 08:53:06 crc kubenswrapper[4873]: I1201 08:53:06.385345 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-767c58684d-z249t"] Dec 01 08:53:06 crc kubenswrapper[4873]: I1201 08:53:06.393435 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-586dc797b5-msvs2" Dec 01 08:53:06 crc kubenswrapper[4873]: I1201 08:53:06.496081 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c9747d5c-eed2-4d46-8cbf-0a01c26624da-apiservice-cert\") pod \"metallb-operator-webhook-server-767c58684d-z249t\" (UID: \"c9747d5c-eed2-4d46-8cbf-0a01c26624da\") " pod="metallb-system/metallb-operator-webhook-server-767c58684d-z249t" Dec 01 08:53:06 crc kubenswrapper[4873]: I1201 08:53:06.496148 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2sn9\" (UniqueName: \"kubernetes.io/projected/c9747d5c-eed2-4d46-8cbf-0a01c26624da-kube-api-access-z2sn9\") pod \"metallb-operator-webhook-server-767c58684d-z249t\" (UID: \"c9747d5c-eed2-4d46-8cbf-0a01c26624da\") " pod="metallb-system/metallb-operator-webhook-server-767c58684d-z249t" Dec 01 08:53:06 crc kubenswrapper[4873]: I1201 08:53:06.496374 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c9747d5c-eed2-4d46-8cbf-0a01c26624da-webhook-cert\") pod \"metallb-operator-webhook-server-767c58684d-z249t\" (UID: \"c9747d5c-eed2-4d46-8cbf-0a01c26624da\") " pod="metallb-system/metallb-operator-webhook-server-767c58684d-z249t" Dec 01 08:53:06 crc kubenswrapper[4873]: I1201 08:53:06.597796 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c9747d5c-eed2-4d46-8cbf-0a01c26624da-webhook-cert\") pod \"metallb-operator-webhook-server-767c58684d-z249t\" (UID: \"c9747d5c-eed2-4d46-8cbf-0a01c26624da\") " pod="metallb-system/metallb-operator-webhook-server-767c58684d-z249t" Dec 01 08:53:06 crc kubenswrapper[4873]: I1201 08:53:06.598193 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c9747d5c-eed2-4d46-8cbf-0a01c26624da-apiservice-cert\") pod \"metallb-operator-webhook-server-767c58684d-z249t\" (UID: \"c9747d5c-eed2-4d46-8cbf-0a01c26624da\") " pod="metallb-system/metallb-operator-webhook-server-767c58684d-z249t" Dec 01 08:53:06 crc kubenswrapper[4873]: I1201 08:53:06.598237 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2sn9\" (UniqueName: \"kubernetes.io/projected/c9747d5c-eed2-4d46-8cbf-0a01c26624da-kube-api-access-z2sn9\") pod \"metallb-operator-webhook-server-767c58684d-z249t\" (UID: \"c9747d5c-eed2-4d46-8cbf-0a01c26624da\") " pod="metallb-system/metallb-operator-webhook-server-767c58684d-z249t" Dec 01 08:53:06 crc kubenswrapper[4873]: I1201 08:53:06.604667 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c9747d5c-eed2-4d46-8cbf-0a01c26624da-webhook-cert\") pod \"metallb-operator-webhook-server-767c58684d-z249t\" (UID: \"c9747d5c-eed2-4d46-8cbf-0a01c26624da\") " pod="metallb-system/metallb-operator-webhook-server-767c58684d-z249t" Dec 01 08:53:06 crc kubenswrapper[4873]: I1201 08:53:06.604694 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c9747d5c-eed2-4d46-8cbf-0a01c26624da-apiservice-cert\") pod \"metallb-operator-webhook-server-767c58684d-z249t\" (UID: \"c9747d5c-eed2-4d46-8cbf-0a01c26624da\") " pod="metallb-system/metallb-operator-webhook-server-767c58684d-z249t" Dec 01 08:53:06 crc kubenswrapper[4873]: I1201 08:53:06.616304 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2sn9\" (UniqueName: \"kubernetes.io/projected/c9747d5c-eed2-4d46-8cbf-0a01c26624da-kube-api-access-z2sn9\") pod \"metallb-operator-webhook-server-767c58684d-z249t\" (UID: \"c9747d5c-eed2-4d46-8cbf-0a01c26624da\") " pod="metallb-system/metallb-operator-webhook-server-767c58684d-z249t" Dec 01 08:53:06 crc kubenswrapper[4873]: I1201 08:53:06.676745 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-767c58684d-z249t" Dec 01 08:53:06 crc kubenswrapper[4873]: I1201 08:53:06.916088 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-586dc797b5-msvs2"] Dec 01 08:53:06 crc kubenswrapper[4873]: W1201 08:53:06.927833 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5b1a0fb8_99da_4202_8c8e_6f5fbcec1908.slice/crio-02b08a5f6ba4ea7ad0db646d9cbad3a560be117c78962d1310f2eda9d5a8da56 WatchSource:0}: Error finding container 02b08a5f6ba4ea7ad0db646d9cbad3a560be117c78962d1310f2eda9d5a8da56: Status 404 returned error can't find the container with id 02b08a5f6ba4ea7ad0db646d9cbad3a560be117c78962d1310f2eda9d5a8da56 Dec 01 08:53:07 crc kubenswrapper[4873]: I1201 08:53:07.031406 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-767c58684d-z249t"] Dec 01 08:53:07 crc kubenswrapper[4873]: W1201 08:53:07.040697 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc9747d5c_eed2_4d46_8cbf_0a01c26624da.slice/crio-d3ca5605e325169233d8f959fe72eac26df37181ea4d1e93a9b95330fd46edd7 WatchSource:0}: Error finding container d3ca5605e325169233d8f959fe72eac26df37181ea4d1e93a9b95330fd46edd7: Status 404 returned error can't find the container with id d3ca5605e325169233d8f959fe72eac26df37181ea4d1e93a9b95330fd46edd7 Dec 01 08:53:07 crc kubenswrapper[4873]: I1201 08:53:07.850578 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-586dc797b5-msvs2" event={"ID":"5b1a0fb8-99da-4202-8c8e-6f5fbcec1908","Type":"ContainerStarted","Data":"02b08a5f6ba4ea7ad0db646d9cbad3a560be117c78962d1310f2eda9d5a8da56"} Dec 01 08:53:07 crc kubenswrapper[4873]: I1201 08:53:07.851624 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-767c58684d-z249t" event={"ID":"c9747d5c-eed2-4d46-8cbf-0a01c26624da","Type":"ContainerStarted","Data":"d3ca5605e325169233d8f959fe72eac26df37181ea4d1e93a9b95330fd46edd7"} Dec 01 08:53:12 crc kubenswrapper[4873]: I1201 08:53:12.896027 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-767c58684d-z249t" event={"ID":"c9747d5c-eed2-4d46-8cbf-0a01c26624da","Type":"ContainerStarted","Data":"9c16f69ec64081c9387a7b101b8a79a8d1b90abb0ca2b520082a894ac9b9ed69"} Dec 01 08:53:12 crc kubenswrapper[4873]: I1201 08:53:12.896767 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-767c58684d-z249t" Dec 01 08:53:12 crc kubenswrapper[4873]: I1201 08:53:12.898276 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-586dc797b5-msvs2" event={"ID":"5b1a0fb8-99da-4202-8c8e-6f5fbcec1908","Type":"ContainerStarted","Data":"f4b885496b862bcaa0b6d838c0953f147ea04fd18a96eb519dc772c33a579480"} Dec 01 08:53:12 crc kubenswrapper[4873]: I1201 08:53:12.898461 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-586dc797b5-msvs2" Dec 01 08:53:12 crc kubenswrapper[4873]: I1201 08:53:12.984960 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-767c58684d-z249t" podStartSLOduration=2.076302135 podStartE2EDuration="6.984931213s" podCreationTimestamp="2025-12-01 08:53:06 +0000 UTC" firstStartedPulling="2025-12-01 08:53:07.044262646 +0000 UTC m=+762.946371185" lastFinishedPulling="2025-12-01 08:53:11.952891724 +0000 UTC m=+767.855000263" observedRunningTime="2025-12-01 08:53:12.974394685 +0000 UTC m=+768.876503224" watchObservedRunningTime="2025-12-01 08:53:12.984931213 +0000 UTC m=+768.887039752" Dec 01 08:53:13 crc kubenswrapper[4873]: I1201 08:53:13.031167 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-586dc797b5-msvs2" podStartSLOduration=2.089783975 podStartE2EDuration="7.031137944s" podCreationTimestamp="2025-12-01 08:53:06 +0000 UTC" firstStartedPulling="2025-12-01 08:53:06.930749978 +0000 UTC m=+762.832858517" lastFinishedPulling="2025-12-01 08:53:11.872103957 +0000 UTC m=+767.774212486" observedRunningTime="2025-12-01 08:53:13.023329153 +0000 UTC m=+768.925437692" watchObservedRunningTime="2025-12-01 08:53:13.031137944 +0000 UTC m=+768.933246483" Dec 01 08:53:26 crc kubenswrapper[4873]: I1201 08:53:26.685035 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-767c58684d-z249t" Dec 01 08:53:46 crc kubenswrapper[4873]: I1201 08:53:46.396908 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-586dc797b5-msvs2" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.129271 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-v584n"] Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.131686 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-v584n" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.134313 4873 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.134437 4873 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-j2m7k" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.134793 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.148763 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-ffvzp"] Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.150399 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-ffvzp" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.151721 4873 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.158848 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-ffvzp"] Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.241532 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-cbkfq"] Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.242640 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-cbkfq" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.245364 4873 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.246240 4873 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-5hkp4" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.246914 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.247031 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s57tg\" (UniqueName: \"kubernetes.io/projected/065b9940-14f0-45e2-9d92-3e7173fce9b2-kube-api-access-s57tg\") pod \"frr-k8s-v584n\" (UID: \"065b9940-14f0-45e2-9d92-3e7173fce9b2\") " pod="metallb-system/frr-k8s-v584n" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.247094 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/065b9940-14f0-45e2-9d92-3e7173fce9b2-metrics-certs\") pod \"frr-k8s-v584n\" (UID: \"065b9940-14f0-45e2-9d92-3e7173fce9b2\") " pod="metallb-system/frr-k8s-v584n" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.247350 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/065b9940-14f0-45e2-9d92-3e7173fce9b2-frr-conf\") pod \"frr-k8s-v584n\" (UID: \"065b9940-14f0-45e2-9d92-3e7173fce9b2\") " pod="metallb-system/frr-k8s-v584n" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.247570 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/065b9940-14f0-45e2-9d92-3e7173fce9b2-frr-sockets\") pod \"frr-k8s-v584n\" (UID: \"065b9940-14f0-45e2-9d92-3e7173fce9b2\") " pod="metallb-system/frr-k8s-v584n" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.247713 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/065b9940-14f0-45e2-9d92-3e7173fce9b2-reloader\") pod \"frr-k8s-v584n\" (UID: \"065b9940-14f0-45e2-9d92-3e7173fce9b2\") " pod="metallb-system/frr-k8s-v584n" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.247878 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/31ad102b-37f9-491d-b017-20c6a3bc4973-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-ffvzp\" (UID: \"31ad102b-37f9-491d-b017-20c6a3bc4973\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-ffvzp" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.247923 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2mxb\" (UniqueName: \"kubernetes.io/projected/31ad102b-37f9-491d-b017-20c6a3bc4973-kube-api-access-h2mxb\") pod \"frr-k8s-webhook-server-7fcb986d4-ffvzp\" (UID: \"31ad102b-37f9-491d-b017-20c6a3bc4973\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-ffvzp" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.248055 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/065b9940-14f0-45e2-9d92-3e7173fce9b2-frr-startup\") pod \"frr-k8s-v584n\" (UID: \"065b9940-14f0-45e2-9d92-3e7173fce9b2\") " pod="metallb-system/frr-k8s-v584n" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.248404 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/065b9940-14f0-45e2-9d92-3e7173fce9b2-metrics\") pod \"frr-k8s-v584n\" (UID: \"065b9940-14f0-45e2-9d92-3e7173fce9b2\") " pod="metallb-system/frr-k8s-v584n" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.248473 4873 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.261354 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-6wb84"] Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.263351 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-6wb84" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.265950 4873 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.280532 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-6wb84"] Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.351106 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/8f181c2f-7d70-4dd9-bcab-ae125dfa4037-memberlist\") pod \"speaker-cbkfq\" (UID: \"8f181c2f-7d70-4dd9-bcab-ae125dfa4037\") " pod="metallb-system/speaker-cbkfq" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.351207 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/065b9940-14f0-45e2-9d92-3e7173fce9b2-frr-sockets\") pod \"frr-k8s-v584n\" (UID: \"065b9940-14f0-45e2-9d92-3e7173fce9b2\") " pod="metallb-system/frr-k8s-v584n" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.351235 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0bd118a2-14e2-438a-bea0-2cd777f71e5c-metrics-certs\") pod \"controller-f8648f98b-6wb84\" (UID: \"0bd118a2-14e2-438a-bea0-2cd777f71e5c\") " pod="metallb-system/controller-f8648f98b-6wb84" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.351273 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/065b9940-14f0-45e2-9d92-3e7173fce9b2-reloader\") pod \"frr-k8s-v584n\" (UID: \"065b9940-14f0-45e2-9d92-3e7173fce9b2\") " pod="metallb-system/frr-k8s-v584n" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.351298 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zd5kc\" (UniqueName: \"kubernetes.io/projected/0bd118a2-14e2-438a-bea0-2cd777f71e5c-kube-api-access-zd5kc\") pod \"controller-f8648f98b-6wb84\" (UID: \"0bd118a2-14e2-438a-bea0-2cd777f71e5c\") " pod="metallb-system/controller-f8648f98b-6wb84" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.351337 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/8f181c2f-7d70-4dd9-bcab-ae125dfa4037-metallb-excludel2\") pod \"speaker-cbkfq\" (UID: \"8f181c2f-7d70-4dd9-bcab-ae125dfa4037\") " pod="metallb-system/speaker-cbkfq" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.351363 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/31ad102b-37f9-491d-b017-20c6a3bc4973-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-ffvzp\" (UID: \"31ad102b-37f9-491d-b017-20c6a3bc4973\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-ffvzp" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.351388 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2mxb\" (UniqueName: \"kubernetes.io/projected/31ad102b-37f9-491d-b017-20c6a3bc4973-kube-api-access-h2mxb\") pod \"frr-k8s-webhook-server-7fcb986d4-ffvzp\" (UID: \"31ad102b-37f9-491d-b017-20c6a3bc4973\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-ffvzp" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.351423 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/065b9940-14f0-45e2-9d92-3e7173fce9b2-frr-startup\") pod \"frr-k8s-v584n\" (UID: \"065b9940-14f0-45e2-9d92-3e7173fce9b2\") " pod="metallb-system/frr-k8s-v584n" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.351461 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/065b9940-14f0-45e2-9d92-3e7173fce9b2-metrics\") pod \"frr-k8s-v584n\" (UID: \"065b9940-14f0-45e2-9d92-3e7173fce9b2\") " pod="metallb-system/frr-k8s-v584n" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.351486 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s57tg\" (UniqueName: \"kubernetes.io/projected/065b9940-14f0-45e2-9d92-3e7173fce9b2-kube-api-access-s57tg\") pod \"frr-k8s-v584n\" (UID: \"065b9940-14f0-45e2-9d92-3e7173fce9b2\") " pod="metallb-system/frr-k8s-v584n" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.351517 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f181c2f-7d70-4dd9-bcab-ae125dfa4037-metrics-certs\") pod \"speaker-cbkfq\" (UID: \"8f181c2f-7d70-4dd9-bcab-ae125dfa4037\") " pod="metallb-system/speaker-cbkfq" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.351539 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/065b9940-14f0-45e2-9d92-3e7173fce9b2-metrics-certs\") pod \"frr-k8s-v584n\" (UID: \"065b9940-14f0-45e2-9d92-3e7173fce9b2\") " pod="metallb-system/frr-k8s-v584n" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.351567 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0bd118a2-14e2-438a-bea0-2cd777f71e5c-cert\") pod \"controller-f8648f98b-6wb84\" (UID: \"0bd118a2-14e2-438a-bea0-2cd777f71e5c\") " pod="metallb-system/controller-f8648f98b-6wb84" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.351598 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/065b9940-14f0-45e2-9d92-3e7173fce9b2-frr-conf\") pod \"frr-k8s-v584n\" (UID: \"065b9940-14f0-45e2-9d92-3e7173fce9b2\") " pod="metallb-system/frr-k8s-v584n" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.351624 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qq6r8\" (UniqueName: \"kubernetes.io/projected/8f181c2f-7d70-4dd9-bcab-ae125dfa4037-kube-api-access-qq6r8\") pod \"speaker-cbkfq\" (UID: \"8f181c2f-7d70-4dd9-bcab-ae125dfa4037\") " pod="metallb-system/speaker-cbkfq" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.352321 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/065b9940-14f0-45e2-9d92-3e7173fce9b2-frr-sockets\") pod \"frr-k8s-v584n\" (UID: \"065b9940-14f0-45e2-9d92-3e7173fce9b2\") " pod="metallb-system/frr-k8s-v584n" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.352621 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/065b9940-14f0-45e2-9d92-3e7173fce9b2-metrics\") pod \"frr-k8s-v584n\" (UID: \"065b9940-14f0-45e2-9d92-3e7173fce9b2\") " pod="metallb-system/frr-k8s-v584n" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.352956 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/065b9940-14f0-45e2-9d92-3e7173fce9b2-reloader\") pod \"frr-k8s-v584n\" (UID: \"065b9940-14f0-45e2-9d92-3e7173fce9b2\") " pod="metallb-system/frr-k8s-v584n" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.353352 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/065b9940-14f0-45e2-9d92-3e7173fce9b2-frr-conf\") pod \"frr-k8s-v584n\" (UID: \"065b9940-14f0-45e2-9d92-3e7173fce9b2\") " pod="metallb-system/frr-k8s-v584n" Dec 01 08:53:47 crc kubenswrapper[4873]: E1201 08:53:47.353446 4873 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Dec 01 08:53:47 crc kubenswrapper[4873]: E1201 08:53:47.353503 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/31ad102b-37f9-491d-b017-20c6a3bc4973-cert podName:31ad102b-37f9-491d-b017-20c6a3bc4973 nodeName:}" failed. No retries permitted until 2025-12-01 08:53:47.853479292 +0000 UTC m=+803.755587831 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/31ad102b-37f9-491d-b017-20c6a3bc4973-cert") pod "frr-k8s-webhook-server-7fcb986d4-ffvzp" (UID: "31ad102b-37f9-491d-b017-20c6a3bc4973") : secret "frr-k8s-webhook-server-cert" not found Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.353505 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/065b9940-14f0-45e2-9d92-3e7173fce9b2-frr-startup\") pod \"frr-k8s-v584n\" (UID: \"065b9940-14f0-45e2-9d92-3e7173fce9b2\") " pod="metallb-system/frr-k8s-v584n" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.390097 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s57tg\" (UniqueName: \"kubernetes.io/projected/065b9940-14f0-45e2-9d92-3e7173fce9b2-kube-api-access-s57tg\") pod \"frr-k8s-v584n\" (UID: \"065b9940-14f0-45e2-9d92-3e7173fce9b2\") " pod="metallb-system/frr-k8s-v584n" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.394836 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/065b9940-14f0-45e2-9d92-3e7173fce9b2-metrics-certs\") pod \"frr-k8s-v584n\" (UID: \"065b9940-14f0-45e2-9d92-3e7173fce9b2\") " pod="metallb-system/frr-k8s-v584n" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.417168 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2mxb\" (UniqueName: \"kubernetes.io/projected/31ad102b-37f9-491d-b017-20c6a3bc4973-kube-api-access-h2mxb\") pod \"frr-k8s-webhook-server-7fcb986d4-ffvzp\" (UID: \"31ad102b-37f9-491d-b017-20c6a3bc4973\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-ffvzp" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.450513 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-v584n" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.452511 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f181c2f-7d70-4dd9-bcab-ae125dfa4037-metrics-certs\") pod \"speaker-cbkfq\" (UID: \"8f181c2f-7d70-4dd9-bcab-ae125dfa4037\") " pod="metallb-system/speaker-cbkfq" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.452556 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0bd118a2-14e2-438a-bea0-2cd777f71e5c-cert\") pod \"controller-f8648f98b-6wb84\" (UID: \"0bd118a2-14e2-438a-bea0-2cd777f71e5c\") " pod="metallb-system/controller-f8648f98b-6wb84" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.452587 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qq6r8\" (UniqueName: \"kubernetes.io/projected/8f181c2f-7d70-4dd9-bcab-ae125dfa4037-kube-api-access-qq6r8\") pod \"speaker-cbkfq\" (UID: \"8f181c2f-7d70-4dd9-bcab-ae125dfa4037\") " pod="metallb-system/speaker-cbkfq" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.452630 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/8f181c2f-7d70-4dd9-bcab-ae125dfa4037-memberlist\") pod \"speaker-cbkfq\" (UID: \"8f181c2f-7d70-4dd9-bcab-ae125dfa4037\") " pod="metallb-system/speaker-cbkfq" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.452654 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0bd118a2-14e2-438a-bea0-2cd777f71e5c-metrics-certs\") pod \"controller-f8648f98b-6wb84\" (UID: \"0bd118a2-14e2-438a-bea0-2cd777f71e5c\") " pod="metallb-system/controller-f8648f98b-6wb84" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.452694 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zd5kc\" (UniqueName: \"kubernetes.io/projected/0bd118a2-14e2-438a-bea0-2cd777f71e5c-kube-api-access-zd5kc\") pod \"controller-f8648f98b-6wb84\" (UID: \"0bd118a2-14e2-438a-bea0-2cd777f71e5c\") " pod="metallb-system/controller-f8648f98b-6wb84" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.452733 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/8f181c2f-7d70-4dd9-bcab-ae125dfa4037-metallb-excludel2\") pod \"speaker-cbkfq\" (UID: \"8f181c2f-7d70-4dd9-bcab-ae125dfa4037\") " pod="metallb-system/speaker-cbkfq" Dec 01 08:53:47 crc kubenswrapper[4873]: E1201 08:53:47.453346 4873 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 01 08:53:47 crc kubenswrapper[4873]: E1201 08:53:47.453432 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f181c2f-7d70-4dd9-bcab-ae125dfa4037-memberlist podName:8f181c2f-7d70-4dd9-bcab-ae125dfa4037 nodeName:}" failed. No retries permitted until 2025-12-01 08:53:47.953404862 +0000 UTC m=+803.855513401 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/8f181c2f-7d70-4dd9-bcab-ae125dfa4037-memberlist") pod "speaker-cbkfq" (UID: "8f181c2f-7d70-4dd9-bcab-ae125dfa4037") : secret "metallb-memberlist" not found Dec 01 08:53:47 crc kubenswrapper[4873]: E1201 08:53:47.453981 4873 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Dec 01 08:53:47 crc kubenswrapper[4873]: E1201 08:53:47.454007 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0bd118a2-14e2-438a-bea0-2cd777f71e5c-metrics-certs podName:0bd118a2-14e2-438a-bea0-2cd777f71e5c nodeName:}" failed. No retries permitted until 2025-12-01 08:53:47.953999227 +0000 UTC m=+803.856107766 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0bd118a2-14e2-438a-bea0-2cd777f71e5c-metrics-certs") pod "controller-f8648f98b-6wb84" (UID: "0bd118a2-14e2-438a-bea0-2cd777f71e5c") : secret "controller-certs-secret" not found Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.454403 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/8f181c2f-7d70-4dd9-bcab-ae125dfa4037-metallb-excludel2\") pod \"speaker-cbkfq\" (UID: \"8f181c2f-7d70-4dd9-bcab-ae125dfa4037\") " pod="metallb-system/speaker-cbkfq" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.458246 4873 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.468556 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f181c2f-7d70-4dd9-bcab-ae125dfa4037-metrics-certs\") pod \"speaker-cbkfq\" (UID: \"8f181c2f-7d70-4dd9-bcab-ae125dfa4037\") " pod="metallb-system/speaker-cbkfq" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.475644 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0bd118a2-14e2-438a-bea0-2cd777f71e5c-cert\") pod \"controller-f8648f98b-6wb84\" (UID: \"0bd118a2-14e2-438a-bea0-2cd777f71e5c\") " pod="metallb-system/controller-f8648f98b-6wb84" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.479219 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zd5kc\" (UniqueName: \"kubernetes.io/projected/0bd118a2-14e2-438a-bea0-2cd777f71e5c-kube-api-access-zd5kc\") pod \"controller-f8648f98b-6wb84\" (UID: \"0bd118a2-14e2-438a-bea0-2cd777f71e5c\") " pod="metallb-system/controller-f8648f98b-6wb84" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.479668 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qq6r8\" (UniqueName: \"kubernetes.io/projected/8f181c2f-7d70-4dd9-bcab-ae125dfa4037-kube-api-access-qq6r8\") pod \"speaker-cbkfq\" (UID: \"8f181c2f-7d70-4dd9-bcab-ae125dfa4037\") " pod="metallb-system/speaker-cbkfq" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.859587 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/31ad102b-37f9-491d-b017-20c6a3bc4973-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-ffvzp\" (UID: \"31ad102b-37f9-491d-b017-20c6a3bc4973\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-ffvzp" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.866152 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/31ad102b-37f9-491d-b017-20c6a3bc4973-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-ffvzp\" (UID: \"31ad102b-37f9-491d-b017-20c6a3bc4973\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-ffvzp" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.961197 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/8f181c2f-7d70-4dd9-bcab-ae125dfa4037-memberlist\") pod \"speaker-cbkfq\" (UID: \"8f181c2f-7d70-4dd9-bcab-ae125dfa4037\") " pod="metallb-system/speaker-cbkfq" Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.961266 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0bd118a2-14e2-438a-bea0-2cd777f71e5c-metrics-certs\") pod \"controller-f8648f98b-6wb84\" (UID: \"0bd118a2-14e2-438a-bea0-2cd777f71e5c\") " pod="metallb-system/controller-f8648f98b-6wb84" Dec 01 08:53:47 crc kubenswrapper[4873]: E1201 08:53:47.961507 4873 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 01 08:53:47 crc kubenswrapper[4873]: E1201 08:53:47.961690 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f181c2f-7d70-4dd9-bcab-ae125dfa4037-memberlist podName:8f181c2f-7d70-4dd9-bcab-ae125dfa4037 nodeName:}" failed. No retries permitted until 2025-12-01 08:53:48.961645139 +0000 UTC m=+804.863753718 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/8f181c2f-7d70-4dd9-bcab-ae125dfa4037-memberlist") pod "speaker-cbkfq" (UID: "8f181c2f-7d70-4dd9-bcab-ae125dfa4037") : secret "metallb-memberlist" not found Dec 01 08:53:47 crc kubenswrapper[4873]: I1201 08:53:47.968711 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0bd118a2-14e2-438a-bea0-2cd777f71e5c-metrics-certs\") pod \"controller-f8648f98b-6wb84\" (UID: \"0bd118a2-14e2-438a-bea0-2cd777f71e5c\") " pod="metallb-system/controller-f8648f98b-6wb84" Dec 01 08:53:48 crc kubenswrapper[4873]: I1201 08:53:48.064457 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-ffvzp" Dec 01 08:53:48 crc kubenswrapper[4873]: I1201 08:53:48.206603 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-6wb84" Dec 01 08:53:48 crc kubenswrapper[4873]: I1201 08:53:48.491266 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-ffvzp"] Dec 01 08:53:48 crc kubenswrapper[4873]: W1201 08:53:48.531296 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0bd118a2_14e2_438a_bea0_2cd777f71e5c.slice/crio-576c27d4419dcbac9f5d811c6e1aec74d6e19e0f3df018e70303c4c331f43b84 WatchSource:0}: Error finding container 576c27d4419dcbac9f5d811c6e1aec74d6e19e0f3df018e70303c4c331f43b84: Status 404 returned error can't find the container with id 576c27d4419dcbac9f5d811c6e1aec74d6e19e0f3df018e70303c4c331f43b84 Dec 01 08:53:48 crc kubenswrapper[4873]: I1201 08:53:48.533903 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-6wb84"] Dec 01 08:53:48 crc kubenswrapper[4873]: I1201 08:53:48.977487 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/8f181c2f-7d70-4dd9-bcab-ae125dfa4037-memberlist\") pod \"speaker-cbkfq\" (UID: \"8f181c2f-7d70-4dd9-bcab-ae125dfa4037\") " pod="metallb-system/speaker-cbkfq" Dec 01 08:53:48 crc kubenswrapper[4873]: I1201 08:53:48.988619 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/8f181c2f-7d70-4dd9-bcab-ae125dfa4037-memberlist\") pod \"speaker-cbkfq\" (UID: \"8f181c2f-7d70-4dd9-bcab-ae125dfa4037\") " pod="metallb-system/speaker-cbkfq" Dec 01 08:53:49 crc kubenswrapper[4873]: I1201 08:53:49.058345 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-cbkfq" Dec 01 08:53:49 crc kubenswrapper[4873]: W1201 08:53:49.084750 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f181c2f_7d70_4dd9_bcab_ae125dfa4037.slice/crio-44e68d1f3456fd2fd37ae37d2b16bbc404495481c5a5f489266c28557083e829 WatchSource:0}: Error finding container 44e68d1f3456fd2fd37ae37d2b16bbc404495481c5a5f489266c28557083e829: Status 404 returned error can't find the container with id 44e68d1f3456fd2fd37ae37d2b16bbc404495481c5a5f489266c28557083e829 Dec 01 08:53:49 crc kubenswrapper[4873]: I1201 08:53:49.126719 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-cbkfq" event={"ID":"8f181c2f-7d70-4dd9-bcab-ae125dfa4037","Type":"ContainerStarted","Data":"44e68d1f3456fd2fd37ae37d2b16bbc404495481c5a5f489266c28557083e829"} Dec 01 08:53:49 crc kubenswrapper[4873]: I1201 08:53:49.128742 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-v584n" event={"ID":"065b9940-14f0-45e2-9d92-3e7173fce9b2","Type":"ContainerStarted","Data":"515a7fd918ba15d9d9681f68993ac100727fff000be84f1f05fc25f4b26ddfc4"} Dec 01 08:53:49 crc kubenswrapper[4873]: I1201 08:53:49.135942 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-6wb84" event={"ID":"0bd118a2-14e2-438a-bea0-2cd777f71e5c","Type":"ContainerStarted","Data":"77c0910a15363f2053c07e9ac43befa900e009ce6e2e0eaa4829f95daa48827c"} Dec 01 08:53:49 crc kubenswrapper[4873]: I1201 08:53:49.136074 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-6wb84" event={"ID":"0bd118a2-14e2-438a-bea0-2cd777f71e5c","Type":"ContainerStarted","Data":"ad6541de104bb3802ed94128388d0d64f1d1ec8634fd871366736de4b0931206"} Dec 01 08:53:49 crc kubenswrapper[4873]: I1201 08:53:49.136115 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-6wb84" event={"ID":"0bd118a2-14e2-438a-bea0-2cd777f71e5c","Type":"ContainerStarted","Data":"576c27d4419dcbac9f5d811c6e1aec74d6e19e0f3df018e70303c4c331f43b84"} Dec 01 08:53:49 crc kubenswrapper[4873]: I1201 08:53:49.136301 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-6wb84" Dec 01 08:53:49 crc kubenswrapper[4873]: I1201 08:53:49.138227 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-ffvzp" event={"ID":"31ad102b-37f9-491d-b017-20c6a3bc4973","Type":"ContainerStarted","Data":"01054d9b5fe1b8b7d3b0e930a95f0f3c0aa2aea12dd22af4d53474537d5f5fe2"} Dec 01 08:53:50 crc kubenswrapper[4873]: I1201 08:53:50.193261 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-cbkfq" event={"ID":"8f181c2f-7d70-4dd9-bcab-ae125dfa4037","Type":"ContainerStarted","Data":"d39745e3d3539c5bca20bf64c6e87b3ccf8d87f49c40b5a02fcb4f3ad7a8e1ac"} Dec 01 08:53:50 crc kubenswrapper[4873]: I1201 08:53:50.193766 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-cbkfq" event={"ID":"8f181c2f-7d70-4dd9-bcab-ae125dfa4037","Type":"ContainerStarted","Data":"4af684259970e78127f8feaad06bea49995a8c56c0fb065ffbf7c3f2a2923d36"} Dec 01 08:53:50 crc kubenswrapper[4873]: I1201 08:53:50.193788 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-cbkfq" Dec 01 08:53:50 crc kubenswrapper[4873]: I1201 08:53:50.213690 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-cbkfq" podStartSLOduration=3.213668547 podStartE2EDuration="3.213668547s" podCreationTimestamp="2025-12-01 08:53:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:53:50.211741529 +0000 UTC m=+806.113850068" watchObservedRunningTime="2025-12-01 08:53:50.213668547 +0000 UTC m=+806.115777086" Dec 01 08:53:50 crc kubenswrapper[4873]: I1201 08:53:50.214576 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-6wb84" podStartSLOduration=3.214568719 podStartE2EDuration="3.214568719s" podCreationTimestamp="2025-12-01 08:53:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:53:49.160585528 +0000 UTC m=+805.062694107" watchObservedRunningTime="2025-12-01 08:53:50.214568719 +0000 UTC m=+806.116677258" Dec 01 08:53:58 crc kubenswrapper[4873]: I1201 08:53:58.219658 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-6wb84" Dec 01 08:53:59 crc kubenswrapper[4873]: I1201 08:53:59.066726 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-cbkfq" Dec 01 08:54:00 crc kubenswrapper[4873]: I1201 08:54:00.505992 4873 generic.go:334] "Generic (PLEG): container finished" podID="065b9940-14f0-45e2-9d92-3e7173fce9b2" containerID="94f08dea8e44a0eee4912d25d20b4caecf88700ac1aef221db744877d3bdef41" exitCode=0 Dec 01 08:54:00 crc kubenswrapper[4873]: I1201 08:54:00.506070 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-v584n" event={"ID":"065b9940-14f0-45e2-9d92-3e7173fce9b2","Type":"ContainerDied","Data":"94f08dea8e44a0eee4912d25d20b4caecf88700ac1aef221db744877d3bdef41"} Dec 01 08:54:00 crc kubenswrapper[4873]: I1201 08:54:00.509115 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-ffvzp" event={"ID":"31ad102b-37f9-491d-b017-20c6a3bc4973","Type":"ContainerStarted","Data":"07a07abc987219b1ff60899ba0e901dfe1e613a1e6fb1271f23406fb3bca6b93"} Dec 01 08:54:00 crc kubenswrapper[4873]: I1201 08:54:00.509389 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-ffvzp" Dec 01 08:54:00 crc kubenswrapper[4873]: I1201 08:54:00.609757 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-ffvzp" podStartSLOduration=2.348871069 podStartE2EDuration="13.609690903s" podCreationTimestamp="2025-12-01 08:53:47 +0000 UTC" firstStartedPulling="2025-12-01 08:53:48.506060411 +0000 UTC m=+804.408168950" lastFinishedPulling="2025-12-01 08:53:59.766880245 +0000 UTC m=+815.668988784" observedRunningTime="2025-12-01 08:54:00.603879607 +0000 UTC m=+816.505988176" watchObservedRunningTime="2025-12-01 08:54:00.609690903 +0000 UTC m=+816.511799452" Dec 01 08:54:01 crc kubenswrapper[4873]: I1201 08:54:01.522490 4873 generic.go:334] "Generic (PLEG): container finished" podID="065b9940-14f0-45e2-9d92-3e7173fce9b2" containerID="5fc0f5bf795317ddb87535c7c5426085e59696c237946718dcfdd9d48f21b0c0" exitCode=0 Dec 01 08:54:01 crc kubenswrapper[4873]: I1201 08:54:01.522563 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-v584n" event={"ID":"065b9940-14f0-45e2-9d92-3e7173fce9b2","Type":"ContainerDied","Data":"5fc0f5bf795317ddb87535c7c5426085e59696c237946718dcfdd9d48f21b0c0"} Dec 01 08:54:02 crc kubenswrapper[4873]: I1201 08:54:02.175293 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-z5nzm"] Dec 01 08:54:02 crc kubenswrapper[4873]: I1201 08:54:02.176879 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-z5nzm" Dec 01 08:54:02 crc kubenswrapper[4873]: I1201 08:54:02.178922 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 01 08:54:02 crc kubenswrapper[4873]: I1201 08:54:02.181134 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-77w75" Dec 01 08:54:02 crc kubenswrapper[4873]: I1201 08:54:02.181162 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 01 08:54:02 crc kubenswrapper[4873]: I1201 08:54:02.188289 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-z5nzm"] Dec 01 08:54:02 crc kubenswrapper[4873]: I1201 08:54:02.295245 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hlhn\" (UniqueName: \"kubernetes.io/projected/5bd51a83-060f-40c3-93d6-5cb871f4bfa7-kube-api-access-7hlhn\") pod \"openstack-operator-index-z5nzm\" (UID: \"5bd51a83-060f-40c3-93d6-5cb871f4bfa7\") " pod="openstack-operators/openstack-operator-index-z5nzm" Dec 01 08:54:02 crc kubenswrapper[4873]: I1201 08:54:02.397184 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hlhn\" (UniqueName: \"kubernetes.io/projected/5bd51a83-060f-40c3-93d6-5cb871f4bfa7-kube-api-access-7hlhn\") pod \"openstack-operator-index-z5nzm\" (UID: \"5bd51a83-060f-40c3-93d6-5cb871f4bfa7\") " pod="openstack-operators/openstack-operator-index-z5nzm" Dec 01 08:54:02 crc kubenswrapper[4873]: I1201 08:54:02.418706 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hlhn\" (UniqueName: \"kubernetes.io/projected/5bd51a83-060f-40c3-93d6-5cb871f4bfa7-kube-api-access-7hlhn\") pod \"openstack-operator-index-z5nzm\" (UID: \"5bd51a83-060f-40c3-93d6-5cb871f4bfa7\") " pod="openstack-operators/openstack-operator-index-z5nzm" Dec 01 08:54:02 crc kubenswrapper[4873]: I1201 08:54:02.496630 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-z5nzm" Dec 01 08:54:02 crc kubenswrapper[4873]: I1201 08:54:02.549786 4873 generic.go:334] "Generic (PLEG): container finished" podID="065b9940-14f0-45e2-9d92-3e7173fce9b2" containerID="810dbfbd9a8c989712e7a966e57682816d08e5aac114588fe1f8efe3b67a49f8" exitCode=0 Dec 01 08:54:02 crc kubenswrapper[4873]: I1201 08:54:02.549892 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-v584n" event={"ID":"065b9940-14f0-45e2-9d92-3e7173fce9b2","Type":"ContainerDied","Data":"810dbfbd9a8c989712e7a966e57682816d08e5aac114588fe1f8efe3b67a49f8"} Dec 01 08:54:02 crc kubenswrapper[4873]: I1201 08:54:02.811081 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-z5nzm"] Dec 01 08:54:02 crc kubenswrapper[4873]: W1201 08:54:02.823847 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5bd51a83_060f_40c3_93d6_5cb871f4bfa7.slice/crio-14c29ffc23667dc953981c44998d5f0828141376471d5c259f802490b4e30636 WatchSource:0}: Error finding container 14c29ffc23667dc953981c44998d5f0828141376471d5c259f802490b4e30636: Status 404 returned error can't find the container with id 14c29ffc23667dc953981c44998d5f0828141376471d5c259f802490b4e30636 Dec 01 08:54:03 crc kubenswrapper[4873]: I1201 08:54:03.564989 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-z5nzm" event={"ID":"5bd51a83-060f-40c3-93d6-5cb871f4bfa7","Type":"ContainerStarted","Data":"14c29ffc23667dc953981c44998d5f0828141376471d5c259f802490b4e30636"} Dec 01 08:54:03 crc kubenswrapper[4873]: I1201 08:54:03.576104 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-v584n" event={"ID":"065b9940-14f0-45e2-9d92-3e7173fce9b2","Type":"ContainerStarted","Data":"93df250f8dc1f1546ae5f24101ee23d299c2ba16cc53bcdba23d7f04dc5631f2"} Dec 01 08:54:03 crc kubenswrapper[4873]: I1201 08:54:03.576174 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-v584n" event={"ID":"065b9940-14f0-45e2-9d92-3e7173fce9b2","Type":"ContainerStarted","Data":"dcc8dead2964b4222e3035afc80d781c6d6cecad51c1dde6dabda0afc6b51f89"} Dec 01 08:54:03 crc kubenswrapper[4873]: I1201 08:54:03.576191 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-v584n" event={"ID":"065b9940-14f0-45e2-9d92-3e7173fce9b2","Type":"ContainerStarted","Data":"624fcae30ad6152a04f33f861aeee15ef240191bcfadfdc4e023d00d68c9390a"} Dec 01 08:54:03 crc kubenswrapper[4873]: I1201 08:54:03.576203 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-v584n" event={"ID":"065b9940-14f0-45e2-9d92-3e7173fce9b2","Type":"ContainerStarted","Data":"a13b7e971b60a212b1c36427f282ffdc181fb5f2df2e7c2f06b4f036986d3114"} Dec 01 08:54:04 crc kubenswrapper[4873]: I1201 08:54:04.590162 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-v584n" event={"ID":"065b9940-14f0-45e2-9d92-3e7173fce9b2","Type":"ContainerStarted","Data":"afee5025e82de6286c17fc7bb17d27165cd8ca2d9abc1e8533ba837bbaa4ddaa"} Dec 01 08:54:05 crc kubenswrapper[4873]: I1201 08:54:05.549259 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-z5nzm"] Dec 01 08:54:06 crc kubenswrapper[4873]: I1201 08:54:06.159063 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-mqm6p"] Dec 01 08:54:06 crc kubenswrapper[4873]: I1201 08:54:06.160845 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-mqm6p" Dec 01 08:54:06 crc kubenswrapper[4873]: I1201 08:54:06.171410 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-mqm6p"] Dec 01 08:54:06 crc kubenswrapper[4873]: I1201 08:54:06.258548 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9kqb\" (UniqueName: \"kubernetes.io/projected/b1309d4c-45e2-444c-afc5-56aa75b9abf5-kube-api-access-p9kqb\") pod \"openstack-operator-index-mqm6p\" (UID: \"b1309d4c-45e2-444c-afc5-56aa75b9abf5\") " pod="openstack-operators/openstack-operator-index-mqm6p" Dec 01 08:54:06 crc kubenswrapper[4873]: I1201 08:54:06.359603 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9kqb\" (UniqueName: \"kubernetes.io/projected/b1309d4c-45e2-444c-afc5-56aa75b9abf5-kube-api-access-p9kqb\") pod \"openstack-operator-index-mqm6p\" (UID: \"b1309d4c-45e2-444c-afc5-56aa75b9abf5\") " pod="openstack-operators/openstack-operator-index-mqm6p" Dec 01 08:54:06 crc kubenswrapper[4873]: I1201 08:54:06.386528 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9kqb\" (UniqueName: \"kubernetes.io/projected/b1309d4c-45e2-444c-afc5-56aa75b9abf5-kube-api-access-p9kqb\") pod \"openstack-operator-index-mqm6p\" (UID: \"b1309d4c-45e2-444c-afc5-56aa75b9abf5\") " pod="openstack-operators/openstack-operator-index-mqm6p" Dec 01 08:54:06 crc kubenswrapper[4873]: I1201 08:54:06.480059 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-mqm6p" Dec 01 08:54:06 crc kubenswrapper[4873]: I1201 08:54:06.632621 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-v584n" event={"ID":"065b9940-14f0-45e2-9d92-3e7173fce9b2","Type":"ContainerStarted","Data":"89aa0e4bbd0fe06cd9aebee799c9559108ef9c92a4af95345299bd4227ea9626"} Dec 01 08:54:06 crc kubenswrapper[4873]: I1201 08:54:06.633146 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-v584n" Dec 01 08:54:06 crc kubenswrapper[4873]: I1201 08:54:06.637415 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-z5nzm" podUID="5bd51a83-060f-40c3-93d6-5cb871f4bfa7" containerName="registry-server" containerID="cri-o://1ae450ace8125a31eea27d843ad753ecb4909a7b2521d94ddac2630c6c36fba9" gracePeriod=2 Dec 01 08:54:06 crc kubenswrapper[4873]: I1201 08:54:06.637187 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-z5nzm" event={"ID":"5bd51a83-060f-40c3-93d6-5cb871f4bfa7","Type":"ContainerStarted","Data":"1ae450ace8125a31eea27d843ad753ecb4909a7b2521d94ddac2630c6c36fba9"} Dec 01 08:54:06 crc kubenswrapper[4873]: I1201 08:54:06.667159 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-v584n" podStartSLOduration=8.241709582 podStartE2EDuration="19.667133564s" podCreationTimestamp="2025-12-01 08:53:47 +0000 UTC" firstStartedPulling="2025-12-01 08:53:48.317540344 +0000 UTC m=+804.219648873" lastFinishedPulling="2025-12-01 08:53:59.742964316 +0000 UTC m=+815.645072855" observedRunningTime="2025-12-01 08:54:06.658255262 +0000 UTC m=+822.560363811" watchObservedRunningTime="2025-12-01 08:54:06.667133564 +0000 UTC m=+822.569242103" Dec 01 08:54:06 crc kubenswrapper[4873]: I1201 08:54:06.688531 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-z5nzm" podStartSLOduration=1.961635741 podStartE2EDuration="4.688503429s" podCreationTimestamp="2025-12-01 08:54:02 +0000 UTC" firstStartedPulling="2025-12-01 08:54:02.82855334 +0000 UTC m=+818.730661879" lastFinishedPulling="2025-12-01 08:54:05.555421018 +0000 UTC m=+821.457529567" observedRunningTime="2025-12-01 08:54:06.68176019 +0000 UTC m=+822.583868729" watchObservedRunningTime="2025-12-01 08:54:06.688503429 +0000 UTC m=+822.590611968" Dec 01 08:54:06 crc kubenswrapper[4873]: I1201 08:54:06.759051 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-mqm6p"] Dec 01 08:54:07 crc kubenswrapper[4873]: I1201 08:54:07.013626 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-z5nzm" Dec 01 08:54:07 crc kubenswrapper[4873]: I1201 08:54:07.070720 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7hlhn\" (UniqueName: \"kubernetes.io/projected/5bd51a83-060f-40c3-93d6-5cb871f4bfa7-kube-api-access-7hlhn\") pod \"5bd51a83-060f-40c3-93d6-5cb871f4bfa7\" (UID: \"5bd51a83-060f-40c3-93d6-5cb871f4bfa7\") " Dec 01 08:54:07 crc kubenswrapper[4873]: I1201 08:54:07.077854 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bd51a83-060f-40c3-93d6-5cb871f4bfa7-kube-api-access-7hlhn" (OuterVolumeSpecName: "kube-api-access-7hlhn") pod "5bd51a83-060f-40c3-93d6-5cb871f4bfa7" (UID: "5bd51a83-060f-40c3-93d6-5cb871f4bfa7"). InnerVolumeSpecName "kube-api-access-7hlhn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:54:07 crc kubenswrapper[4873]: I1201 08:54:07.172725 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7hlhn\" (UniqueName: \"kubernetes.io/projected/5bd51a83-060f-40c3-93d6-5cb871f4bfa7-kube-api-access-7hlhn\") on node \"crc\" DevicePath \"\"" Dec 01 08:54:07 crc kubenswrapper[4873]: I1201 08:54:07.451310 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-v584n" Dec 01 08:54:07 crc kubenswrapper[4873]: I1201 08:54:07.523113 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-v584n" Dec 01 08:54:07 crc kubenswrapper[4873]: I1201 08:54:07.648001 4873 generic.go:334] "Generic (PLEG): container finished" podID="5bd51a83-060f-40c3-93d6-5cb871f4bfa7" containerID="1ae450ace8125a31eea27d843ad753ecb4909a7b2521d94ddac2630c6c36fba9" exitCode=0 Dec 01 08:54:07 crc kubenswrapper[4873]: I1201 08:54:07.648117 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-z5nzm" Dec 01 08:54:07 crc kubenswrapper[4873]: I1201 08:54:07.648167 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-z5nzm" event={"ID":"5bd51a83-060f-40c3-93d6-5cb871f4bfa7","Type":"ContainerDied","Data":"1ae450ace8125a31eea27d843ad753ecb4909a7b2521d94ddac2630c6c36fba9"} Dec 01 08:54:07 crc kubenswrapper[4873]: I1201 08:54:07.648256 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-z5nzm" event={"ID":"5bd51a83-060f-40c3-93d6-5cb871f4bfa7","Type":"ContainerDied","Data":"14c29ffc23667dc953981c44998d5f0828141376471d5c259f802490b4e30636"} Dec 01 08:54:07 crc kubenswrapper[4873]: I1201 08:54:07.648290 4873 scope.go:117] "RemoveContainer" containerID="1ae450ace8125a31eea27d843ad753ecb4909a7b2521d94ddac2630c6c36fba9" Dec 01 08:54:07 crc kubenswrapper[4873]: I1201 08:54:07.652522 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-mqm6p" event={"ID":"b1309d4c-45e2-444c-afc5-56aa75b9abf5","Type":"ContainerStarted","Data":"97239b402b4427e067eacdc68828930507f06d98fbed9b12502cc0a73eb2bf85"} Dec 01 08:54:07 crc kubenswrapper[4873]: I1201 08:54:07.652691 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-mqm6p" event={"ID":"b1309d4c-45e2-444c-afc5-56aa75b9abf5","Type":"ContainerStarted","Data":"9354e0c574bf94ad15326edc6b7e39e6d3868eea0e0e4fb5a6f5ea42c0d5b275"} Dec 01 08:54:07 crc kubenswrapper[4873]: I1201 08:54:07.671979 4873 scope.go:117] "RemoveContainer" containerID="1ae450ace8125a31eea27d843ad753ecb4909a7b2521d94ddac2630c6c36fba9" Dec 01 08:54:07 crc kubenswrapper[4873]: E1201 08:54:07.672979 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ae450ace8125a31eea27d843ad753ecb4909a7b2521d94ddac2630c6c36fba9\": container with ID starting with 1ae450ace8125a31eea27d843ad753ecb4909a7b2521d94ddac2630c6c36fba9 not found: ID does not exist" containerID="1ae450ace8125a31eea27d843ad753ecb4909a7b2521d94ddac2630c6c36fba9" Dec 01 08:54:07 crc kubenswrapper[4873]: I1201 08:54:07.673075 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ae450ace8125a31eea27d843ad753ecb4909a7b2521d94ddac2630c6c36fba9"} err="failed to get container status \"1ae450ace8125a31eea27d843ad753ecb4909a7b2521d94ddac2630c6c36fba9\": rpc error: code = NotFound desc = could not find container \"1ae450ace8125a31eea27d843ad753ecb4909a7b2521d94ddac2630c6c36fba9\": container with ID starting with 1ae450ace8125a31eea27d843ad753ecb4909a7b2521d94ddac2630c6c36fba9 not found: ID does not exist" Dec 01 08:54:07 crc kubenswrapper[4873]: I1201 08:54:07.682724 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-mqm6p" podStartSLOduration=1.636110479 podStartE2EDuration="1.682689154s" podCreationTimestamp="2025-12-01 08:54:06 +0000 UTC" firstStartedPulling="2025-12-01 08:54:06.772228494 +0000 UTC m=+822.674337033" lastFinishedPulling="2025-12-01 08:54:06.818807139 +0000 UTC m=+822.720915708" observedRunningTime="2025-12-01 08:54:07.675258838 +0000 UTC m=+823.577367407" watchObservedRunningTime="2025-12-01 08:54:07.682689154 +0000 UTC m=+823.584797703" Dec 01 08:54:07 crc kubenswrapper[4873]: I1201 08:54:07.744854 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-z5nzm"] Dec 01 08:54:07 crc kubenswrapper[4873]: I1201 08:54:07.752325 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-z5nzm"] Dec 01 08:54:08 crc kubenswrapper[4873]: I1201 08:54:08.440143 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5bd51a83-060f-40c3-93d6-5cb871f4bfa7" path="/var/lib/kubelet/pods/5bd51a83-060f-40c3-93d6-5cb871f4bfa7/volumes" Dec 01 08:54:16 crc kubenswrapper[4873]: I1201 08:54:16.481104 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-mqm6p" Dec 01 08:54:16 crc kubenswrapper[4873]: I1201 08:54:16.482296 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-mqm6p" Dec 01 08:54:16 crc kubenswrapper[4873]: I1201 08:54:16.530206 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-mqm6p" Dec 01 08:54:16 crc kubenswrapper[4873]: I1201 08:54:16.763249 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-mqm6p" Dec 01 08:54:17 crc kubenswrapper[4873]: I1201 08:54:17.456312 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-v584n" Dec 01 08:54:18 crc kubenswrapper[4873]: I1201 08:54:18.074567 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-ffvzp" Dec 01 08:54:24 crc kubenswrapper[4873]: I1201 08:54:24.219868 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/d92c7d608fded6ff7325ca25db46cfa3dc2f9a2773cb154702b56fef16ks4c9"] Dec 01 08:54:24 crc kubenswrapper[4873]: E1201 08:54:24.220263 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bd51a83-060f-40c3-93d6-5cb871f4bfa7" containerName="registry-server" Dec 01 08:54:24 crc kubenswrapper[4873]: I1201 08:54:24.220286 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bd51a83-060f-40c3-93d6-5cb871f4bfa7" containerName="registry-server" Dec 01 08:54:24 crc kubenswrapper[4873]: I1201 08:54:24.220461 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bd51a83-060f-40c3-93d6-5cb871f4bfa7" containerName="registry-server" Dec 01 08:54:24 crc kubenswrapper[4873]: I1201 08:54:24.221669 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d92c7d608fded6ff7325ca25db46cfa3dc2f9a2773cb154702b56fef16ks4c9" Dec 01 08:54:24 crc kubenswrapper[4873]: I1201 08:54:24.225762 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-nwncn" Dec 01 08:54:24 crc kubenswrapper[4873]: I1201 08:54:24.246254 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/d92c7d608fded6ff7325ca25db46cfa3dc2f9a2773cb154702b56fef16ks4c9"] Dec 01 08:54:24 crc kubenswrapper[4873]: I1201 08:54:24.359795 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxhdg\" (UniqueName: \"kubernetes.io/projected/08dfa43f-80d7-4785-a0ce-c8b6d8d95351-kube-api-access-zxhdg\") pod \"d92c7d608fded6ff7325ca25db46cfa3dc2f9a2773cb154702b56fef16ks4c9\" (UID: \"08dfa43f-80d7-4785-a0ce-c8b6d8d95351\") " pod="openstack-operators/d92c7d608fded6ff7325ca25db46cfa3dc2f9a2773cb154702b56fef16ks4c9" Dec 01 08:54:24 crc kubenswrapper[4873]: I1201 08:54:24.359896 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/08dfa43f-80d7-4785-a0ce-c8b6d8d95351-util\") pod \"d92c7d608fded6ff7325ca25db46cfa3dc2f9a2773cb154702b56fef16ks4c9\" (UID: \"08dfa43f-80d7-4785-a0ce-c8b6d8d95351\") " pod="openstack-operators/d92c7d608fded6ff7325ca25db46cfa3dc2f9a2773cb154702b56fef16ks4c9" Dec 01 08:54:24 crc kubenswrapper[4873]: I1201 08:54:24.359937 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/08dfa43f-80d7-4785-a0ce-c8b6d8d95351-bundle\") pod \"d92c7d608fded6ff7325ca25db46cfa3dc2f9a2773cb154702b56fef16ks4c9\" (UID: \"08dfa43f-80d7-4785-a0ce-c8b6d8d95351\") " pod="openstack-operators/d92c7d608fded6ff7325ca25db46cfa3dc2f9a2773cb154702b56fef16ks4c9" Dec 01 08:54:24 crc kubenswrapper[4873]: I1201 08:54:24.461864 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/08dfa43f-80d7-4785-a0ce-c8b6d8d95351-util\") pod \"d92c7d608fded6ff7325ca25db46cfa3dc2f9a2773cb154702b56fef16ks4c9\" (UID: \"08dfa43f-80d7-4785-a0ce-c8b6d8d95351\") " pod="openstack-operators/d92c7d608fded6ff7325ca25db46cfa3dc2f9a2773cb154702b56fef16ks4c9" Dec 01 08:54:24 crc kubenswrapper[4873]: I1201 08:54:24.461916 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/08dfa43f-80d7-4785-a0ce-c8b6d8d95351-bundle\") pod \"d92c7d608fded6ff7325ca25db46cfa3dc2f9a2773cb154702b56fef16ks4c9\" (UID: \"08dfa43f-80d7-4785-a0ce-c8b6d8d95351\") " pod="openstack-operators/d92c7d608fded6ff7325ca25db46cfa3dc2f9a2773cb154702b56fef16ks4c9" Dec 01 08:54:24 crc kubenswrapper[4873]: I1201 08:54:24.461978 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxhdg\" (UniqueName: \"kubernetes.io/projected/08dfa43f-80d7-4785-a0ce-c8b6d8d95351-kube-api-access-zxhdg\") pod \"d92c7d608fded6ff7325ca25db46cfa3dc2f9a2773cb154702b56fef16ks4c9\" (UID: \"08dfa43f-80d7-4785-a0ce-c8b6d8d95351\") " pod="openstack-operators/d92c7d608fded6ff7325ca25db46cfa3dc2f9a2773cb154702b56fef16ks4c9" Dec 01 08:54:24 crc kubenswrapper[4873]: I1201 08:54:24.462592 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/08dfa43f-80d7-4785-a0ce-c8b6d8d95351-util\") pod \"d92c7d608fded6ff7325ca25db46cfa3dc2f9a2773cb154702b56fef16ks4c9\" (UID: \"08dfa43f-80d7-4785-a0ce-c8b6d8d95351\") " pod="openstack-operators/d92c7d608fded6ff7325ca25db46cfa3dc2f9a2773cb154702b56fef16ks4c9" Dec 01 08:54:24 crc kubenswrapper[4873]: I1201 08:54:24.462658 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/08dfa43f-80d7-4785-a0ce-c8b6d8d95351-bundle\") pod \"d92c7d608fded6ff7325ca25db46cfa3dc2f9a2773cb154702b56fef16ks4c9\" (UID: \"08dfa43f-80d7-4785-a0ce-c8b6d8d95351\") " pod="openstack-operators/d92c7d608fded6ff7325ca25db46cfa3dc2f9a2773cb154702b56fef16ks4c9" Dec 01 08:54:24 crc kubenswrapper[4873]: I1201 08:54:24.491312 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxhdg\" (UniqueName: \"kubernetes.io/projected/08dfa43f-80d7-4785-a0ce-c8b6d8d95351-kube-api-access-zxhdg\") pod \"d92c7d608fded6ff7325ca25db46cfa3dc2f9a2773cb154702b56fef16ks4c9\" (UID: \"08dfa43f-80d7-4785-a0ce-c8b6d8d95351\") " pod="openstack-operators/d92c7d608fded6ff7325ca25db46cfa3dc2f9a2773cb154702b56fef16ks4c9" Dec 01 08:54:24 crc kubenswrapper[4873]: I1201 08:54:24.552738 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-nwncn" Dec 01 08:54:24 crc kubenswrapper[4873]: I1201 08:54:24.561180 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d92c7d608fded6ff7325ca25db46cfa3dc2f9a2773cb154702b56fef16ks4c9" Dec 01 08:54:25 crc kubenswrapper[4873]: I1201 08:54:25.078379 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/d92c7d608fded6ff7325ca25db46cfa3dc2f9a2773cb154702b56fef16ks4c9"] Dec 01 08:54:25 crc kubenswrapper[4873]: W1201 08:54:25.082591 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod08dfa43f_80d7_4785_a0ce_c8b6d8d95351.slice/crio-f28e5e9fd7cb7bde3046b755497699cff3aecb526a12c3ef2716b0c0db3870cc WatchSource:0}: Error finding container f28e5e9fd7cb7bde3046b755497699cff3aecb526a12c3ef2716b0c0db3870cc: Status 404 returned error can't find the container with id f28e5e9fd7cb7bde3046b755497699cff3aecb526a12c3ef2716b0c0db3870cc Dec 01 08:54:25 crc kubenswrapper[4873]: I1201 08:54:25.795644 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d92c7d608fded6ff7325ca25db46cfa3dc2f9a2773cb154702b56fef16ks4c9" event={"ID":"08dfa43f-80d7-4785-a0ce-c8b6d8d95351","Type":"ContainerStarted","Data":"f28e5e9fd7cb7bde3046b755497699cff3aecb526a12c3ef2716b0c0db3870cc"} Dec 01 08:54:26 crc kubenswrapper[4873]: E1201 08:54:26.090755 4873 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod08dfa43f_80d7_4785_a0ce_c8b6d8d95351.slice/crio-conmon-5a6f2b9f65619904e5122d5805963bae466b742b81ee46284a850b29187cf37c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod08dfa43f_80d7_4785_a0ce_c8b6d8d95351.slice/crio-5a6f2b9f65619904e5122d5805963bae466b742b81ee46284a850b29187cf37c.scope\": RecentStats: unable to find data in memory cache]" Dec 01 08:54:26 crc kubenswrapper[4873]: I1201 08:54:26.805346 4873 generic.go:334] "Generic (PLEG): container finished" podID="08dfa43f-80d7-4785-a0ce-c8b6d8d95351" containerID="5a6f2b9f65619904e5122d5805963bae466b742b81ee46284a850b29187cf37c" exitCode=0 Dec 01 08:54:26 crc kubenswrapper[4873]: I1201 08:54:26.805467 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d92c7d608fded6ff7325ca25db46cfa3dc2f9a2773cb154702b56fef16ks4c9" event={"ID":"08dfa43f-80d7-4785-a0ce-c8b6d8d95351","Type":"ContainerDied","Data":"5a6f2b9f65619904e5122d5805963bae466b742b81ee46284a850b29187cf37c"} Dec 01 08:54:27 crc kubenswrapper[4873]: I1201 08:54:27.814923 4873 generic.go:334] "Generic (PLEG): container finished" podID="08dfa43f-80d7-4785-a0ce-c8b6d8d95351" containerID="2f02496a5aca7295cb11c20b498f18c67ebf75002647cbbae810742a3d46f14f" exitCode=0 Dec 01 08:54:27 crc kubenswrapper[4873]: I1201 08:54:27.815056 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d92c7d608fded6ff7325ca25db46cfa3dc2f9a2773cb154702b56fef16ks4c9" event={"ID":"08dfa43f-80d7-4785-a0ce-c8b6d8d95351","Type":"ContainerDied","Data":"2f02496a5aca7295cb11c20b498f18c67ebf75002647cbbae810742a3d46f14f"} Dec 01 08:54:28 crc kubenswrapper[4873]: I1201 08:54:28.826550 4873 generic.go:334] "Generic (PLEG): container finished" podID="08dfa43f-80d7-4785-a0ce-c8b6d8d95351" containerID="778a2e38a5822f3b235065ed2df49e6922f307d08c815aac0c5a20f461632d5e" exitCode=0 Dec 01 08:54:28 crc kubenswrapper[4873]: I1201 08:54:28.826610 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d92c7d608fded6ff7325ca25db46cfa3dc2f9a2773cb154702b56fef16ks4c9" event={"ID":"08dfa43f-80d7-4785-a0ce-c8b6d8d95351","Type":"ContainerDied","Data":"778a2e38a5822f3b235065ed2df49e6922f307d08c815aac0c5a20f461632d5e"} Dec 01 08:54:30 crc kubenswrapper[4873]: I1201 08:54:30.085415 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d92c7d608fded6ff7325ca25db46cfa3dc2f9a2773cb154702b56fef16ks4c9" Dec 01 08:54:30 crc kubenswrapper[4873]: I1201 08:54:30.261455 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/08dfa43f-80d7-4785-a0ce-c8b6d8d95351-bundle\") pod \"08dfa43f-80d7-4785-a0ce-c8b6d8d95351\" (UID: \"08dfa43f-80d7-4785-a0ce-c8b6d8d95351\") " Dec 01 08:54:30 crc kubenswrapper[4873]: I1201 08:54:30.261594 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxhdg\" (UniqueName: \"kubernetes.io/projected/08dfa43f-80d7-4785-a0ce-c8b6d8d95351-kube-api-access-zxhdg\") pod \"08dfa43f-80d7-4785-a0ce-c8b6d8d95351\" (UID: \"08dfa43f-80d7-4785-a0ce-c8b6d8d95351\") " Dec 01 08:54:30 crc kubenswrapper[4873]: I1201 08:54:30.261625 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/08dfa43f-80d7-4785-a0ce-c8b6d8d95351-util\") pod \"08dfa43f-80d7-4785-a0ce-c8b6d8d95351\" (UID: \"08dfa43f-80d7-4785-a0ce-c8b6d8d95351\") " Dec 01 08:54:30 crc kubenswrapper[4873]: I1201 08:54:30.262603 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08dfa43f-80d7-4785-a0ce-c8b6d8d95351-bundle" (OuterVolumeSpecName: "bundle") pod "08dfa43f-80d7-4785-a0ce-c8b6d8d95351" (UID: "08dfa43f-80d7-4785-a0ce-c8b6d8d95351"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:54:30 crc kubenswrapper[4873]: I1201 08:54:30.269048 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08dfa43f-80d7-4785-a0ce-c8b6d8d95351-kube-api-access-zxhdg" (OuterVolumeSpecName: "kube-api-access-zxhdg") pod "08dfa43f-80d7-4785-a0ce-c8b6d8d95351" (UID: "08dfa43f-80d7-4785-a0ce-c8b6d8d95351"). InnerVolumeSpecName "kube-api-access-zxhdg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:54:30 crc kubenswrapper[4873]: I1201 08:54:30.276614 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08dfa43f-80d7-4785-a0ce-c8b6d8d95351-util" (OuterVolumeSpecName: "util") pod "08dfa43f-80d7-4785-a0ce-c8b6d8d95351" (UID: "08dfa43f-80d7-4785-a0ce-c8b6d8d95351"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:54:30 crc kubenswrapper[4873]: I1201 08:54:30.363582 4873 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/08dfa43f-80d7-4785-a0ce-c8b6d8d95351-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:54:30 crc kubenswrapper[4873]: I1201 08:54:30.363638 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxhdg\" (UniqueName: \"kubernetes.io/projected/08dfa43f-80d7-4785-a0ce-c8b6d8d95351-kube-api-access-zxhdg\") on node \"crc\" DevicePath \"\"" Dec 01 08:54:30 crc kubenswrapper[4873]: I1201 08:54:30.363650 4873 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/08dfa43f-80d7-4785-a0ce-c8b6d8d95351-util\") on node \"crc\" DevicePath \"\"" Dec 01 08:54:30 crc kubenswrapper[4873]: I1201 08:54:30.848382 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d92c7d608fded6ff7325ca25db46cfa3dc2f9a2773cb154702b56fef16ks4c9" event={"ID":"08dfa43f-80d7-4785-a0ce-c8b6d8d95351","Type":"ContainerDied","Data":"f28e5e9fd7cb7bde3046b755497699cff3aecb526a12c3ef2716b0c0db3870cc"} Dec 01 08:54:30 crc kubenswrapper[4873]: I1201 08:54:30.848434 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f28e5e9fd7cb7bde3046b755497699cff3aecb526a12c3ef2716b0c0db3870cc" Dec 01 08:54:30 crc kubenswrapper[4873]: I1201 08:54:30.848513 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d92c7d608fded6ff7325ca25db46cfa3dc2f9a2773cb154702b56fef16ks4c9" Dec 01 08:54:37 crc kubenswrapper[4873]: I1201 08:54:37.034993 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7f57cdcc9-jttc7"] Dec 01 08:54:37 crc kubenswrapper[4873]: E1201 08:54:37.036261 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08dfa43f-80d7-4785-a0ce-c8b6d8d95351" containerName="extract" Dec 01 08:54:37 crc kubenswrapper[4873]: I1201 08:54:37.036310 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="08dfa43f-80d7-4785-a0ce-c8b6d8d95351" containerName="extract" Dec 01 08:54:37 crc kubenswrapper[4873]: E1201 08:54:37.036329 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08dfa43f-80d7-4785-a0ce-c8b6d8d95351" containerName="util" Dec 01 08:54:37 crc kubenswrapper[4873]: I1201 08:54:37.036337 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="08dfa43f-80d7-4785-a0ce-c8b6d8d95351" containerName="util" Dec 01 08:54:37 crc kubenswrapper[4873]: E1201 08:54:37.036370 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08dfa43f-80d7-4785-a0ce-c8b6d8d95351" containerName="pull" Dec 01 08:54:37 crc kubenswrapper[4873]: I1201 08:54:37.036378 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="08dfa43f-80d7-4785-a0ce-c8b6d8d95351" containerName="pull" Dec 01 08:54:37 crc kubenswrapper[4873]: I1201 08:54:37.036570 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="08dfa43f-80d7-4785-a0ce-c8b6d8d95351" containerName="extract" Dec 01 08:54:37 crc kubenswrapper[4873]: I1201 08:54:37.037321 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7f57cdcc9-jttc7" Dec 01 08:54:37 crc kubenswrapper[4873]: I1201 08:54:37.040294 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-c7mjg" Dec 01 08:54:37 crc kubenswrapper[4873]: I1201 08:54:37.072042 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tcnfg\" (UniqueName: \"kubernetes.io/projected/c25dac67-107f-43e2-a63e-5843ff31abc8-kube-api-access-tcnfg\") pod \"openstack-operator-controller-operator-7f57cdcc9-jttc7\" (UID: \"c25dac67-107f-43e2-a63e-5843ff31abc8\") " pod="openstack-operators/openstack-operator-controller-operator-7f57cdcc9-jttc7" Dec 01 08:54:37 crc kubenswrapper[4873]: I1201 08:54:37.074666 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7f57cdcc9-jttc7"] Dec 01 08:54:37 crc kubenswrapper[4873]: I1201 08:54:37.173481 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tcnfg\" (UniqueName: \"kubernetes.io/projected/c25dac67-107f-43e2-a63e-5843ff31abc8-kube-api-access-tcnfg\") pod \"openstack-operator-controller-operator-7f57cdcc9-jttc7\" (UID: \"c25dac67-107f-43e2-a63e-5843ff31abc8\") " pod="openstack-operators/openstack-operator-controller-operator-7f57cdcc9-jttc7" Dec 01 08:54:37 crc kubenswrapper[4873]: I1201 08:54:37.197040 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tcnfg\" (UniqueName: \"kubernetes.io/projected/c25dac67-107f-43e2-a63e-5843ff31abc8-kube-api-access-tcnfg\") pod \"openstack-operator-controller-operator-7f57cdcc9-jttc7\" (UID: \"c25dac67-107f-43e2-a63e-5843ff31abc8\") " pod="openstack-operators/openstack-operator-controller-operator-7f57cdcc9-jttc7" Dec 01 08:54:37 crc kubenswrapper[4873]: I1201 08:54:37.358371 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7f57cdcc9-jttc7" Dec 01 08:54:37 crc kubenswrapper[4873]: I1201 08:54:37.668668 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7f57cdcc9-jttc7"] Dec 01 08:54:37 crc kubenswrapper[4873]: I1201 08:54:37.921753 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7f57cdcc9-jttc7" event={"ID":"c25dac67-107f-43e2-a63e-5843ff31abc8","Type":"ContainerStarted","Data":"1a9c2a18bc598fd4072c52cd6453b4065697d21c527baa4a21f7aa1f2ffedde1"} Dec 01 08:54:46 crc kubenswrapper[4873]: I1201 08:54:46.130064 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7f57cdcc9-jttc7" event={"ID":"c25dac67-107f-43e2-a63e-5843ff31abc8","Type":"ContainerStarted","Data":"ac57fa1930294c8da1555a60153ee7dfb045d4ce610278adc0348823ca6d2de7"} Dec 01 08:54:46 crc kubenswrapper[4873]: I1201 08:54:46.130650 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-7f57cdcc9-jttc7" Dec 01 08:54:46 crc kubenswrapper[4873]: I1201 08:54:46.168685 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-7f57cdcc9-jttc7" podStartSLOduration=1.638812122 podStartE2EDuration="9.168656825s" podCreationTimestamp="2025-12-01 08:54:37 +0000 UTC" firstStartedPulling="2025-12-01 08:54:37.675864287 +0000 UTC m=+853.577972846" lastFinishedPulling="2025-12-01 08:54:45.20570901 +0000 UTC m=+861.107817549" observedRunningTime="2025-12-01 08:54:46.163125977 +0000 UTC m=+862.065234516" watchObservedRunningTime="2025-12-01 08:54:46.168656825 +0000 UTC m=+862.070765364" Dec 01 08:54:57 crc kubenswrapper[4873]: I1201 08:54:57.361769 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-7f57cdcc9-jttc7" Dec 01 08:55:01 crc kubenswrapper[4873]: I1201 08:55:01.059674 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:55:01 crc kubenswrapper[4873]: I1201 08:55:01.060151 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.059424 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-5rq8z"] Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.061029 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5rq8z" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.063972 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-qnqdh" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.071762 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-gqm5w"] Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.073153 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-gqm5w" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.076549 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-qshpm" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.084005 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-gqm5w"] Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.092652 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-5rq8z"] Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.100239 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-6r8hb"] Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.102918 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-6r8hb" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.110562 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-q4gwq" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.115648 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-6r8hb"] Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.151961 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shh7b\" (UniqueName: \"kubernetes.io/projected/671ad3cc-9716-42de-b5ae-fb69847d3bd7-kube-api-access-shh7b\") pod \"cinder-operator-controller-manager-859b6ccc6-gqm5w\" (UID: \"671ad3cc-9716-42de-b5ae-fb69847d3bd7\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-gqm5w" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.152028 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nt2h4\" (UniqueName: \"kubernetes.io/projected/723f446d-21bb-432a-b2af-93219593819d-kube-api-access-nt2h4\") pod \"designate-operator-controller-manager-78b4bc895b-6r8hb\" (UID: \"723f446d-21bb-432a-b2af-93219593819d\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-6r8hb" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.152093 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9hnx\" (UniqueName: \"kubernetes.io/projected/51253ba6-e0b5-44ac-8c18-be17b4d13024-kube-api-access-t9hnx\") pod \"barbican-operator-controller-manager-7d9dfd778-5rq8z\" (UID: \"51253ba6-e0b5-44ac-8c18-be17b4d13024\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5rq8z" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.181075 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-5cd474b4ff-wcmv4"] Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.182332 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5cd474b4ff-wcmv4" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.185335 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-x4w4m" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.185810 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-d7c6w"] Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.186864 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-d7c6w" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.194800 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-5v74m" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.210136 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5cd474b4ff-wcmv4"] Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.221561 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-d7c6w"] Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.228819 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-45whz"] Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.252682 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-45whz" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.260310 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-mmpkl" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.262896 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nt2h4\" (UniqueName: \"kubernetes.io/projected/723f446d-21bb-432a-b2af-93219593819d-kube-api-access-nt2h4\") pod \"designate-operator-controller-manager-78b4bc895b-6r8hb\" (UID: \"723f446d-21bb-432a-b2af-93219593819d\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-6r8hb" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.263001 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9hnx\" (UniqueName: \"kubernetes.io/projected/51253ba6-e0b5-44ac-8c18-be17b4d13024-kube-api-access-t9hnx\") pod \"barbican-operator-controller-manager-7d9dfd778-5rq8z\" (UID: \"51253ba6-e0b5-44ac-8c18-be17b4d13024\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5rq8z" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.263831 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shh7b\" (UniqueName: \"kubernetes.io/projected/671ad3cc-9716-42de-b5ae-fb69847d3bd7-kube-api-access-shh7b\") pod \"cinder-operator-controller-manager-859b6ccc6-gqm5w\" (UID: \"671ad3cc-9716-42de-b5ae-fb69847d3bd7\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-gqm5w" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.264099 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-tmdj7"] Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.281577 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-45whz"] Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.281798 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-tmdj7" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.294999 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.295836 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-ptl6k" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.298949 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-tmdj7"] Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.309256 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nt2h4\" (UniqueName: \"kubernetes.io/projected/723f446d-21bb-432a-b2af-93219593819d-kube-api-access-nt2h4\") pod \"designate-operator-controller-manager-78b4bc895b-6r8hb\" (UID: \"723f446d-21bb-432a-b2af-93219593819d\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-6r8hb" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.317478 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9hnx\" (UniqueName: \"kubernetes.io/projected/51253ba6-e0b5-44ac-8c18-be17b4d13024-kube-api-access-t9hnx\") pod \"barbican-operator-controller-manager-7d9dfd778-5rq8z\" (UID: \"51253ba6-e0b5-44ac-8c18-be17b4d13024\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5rq8z" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.367799 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxln9\" (UniqueName: \"kubernetes.io/projected/967215e8-7a18-4507-96c1-4c79c7e1d51a-kube-api-access-xxln9\") pod \"heat-operator-controller-manager-5f64f6f8bb-d7c6w\" (UID: \"967215e8-7a18-4507-96c1-4c79c7e1d51a\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-d7c6w" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.368299 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgqgk\" (UniqueName: \"kubernetes.io/projected/94e7345b-b1ac-46e2-be25-9d64d3d33523-kube-api-access-qgqgk\") pod \"horizon-operator-controller-manager-68c6d99b8f-45whz\" (UID: \"94e7345b-b1ac-46e2-be25-9d64d3d33523\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-45whz" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.368341 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbjxp\" (UniqueName: \"kubernetes.io/projected/f778fd6c-4c8a-4067-b3f2-cb7d98a50bf7-kube-api-access-jbjxp\") pod \"glance-operator-controller-manager-5cd474b4ff-wcmv4\" (UID: \"f778fd6c-4c8a-4067-b3f2-cb7d98a50bf7\") " pod="openstack-operators/glance-operator-controller-manager-5cd474b4ff-wcmv4" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.382824 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shh7b\" (UniqueName: \"kubernetes.io/projected/671ad3cc-9716-42de-b5ae-fb69847d3bd7-kube-api-access-shh7b\") pod \"cinder-operator-controller-manager-859b6ccc6-gqm5w\" (UID: \"671ad3cc-9716-42de-b5ae-fb69847d3bd7\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-gqm5w" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.385568 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-7wmqg"] Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.387639 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7wmqg" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.393937 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-z4t88" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.408641 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-qnqdh" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.408848 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5rq8z" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.422913 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-qshpm" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.433170 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-gqm5w" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.440175 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-q4gwq" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.442219 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-6r8hb" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.471279 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9klvb\" (UniqueName: \"kubernetes.io/projected/4dcc939e-b548-4fb6-814a-30e2aaa8a94a-kube-api-access-9klvb\") pod \"infra-operator-controller-manager-57548d458d-tmdj7\" (UID: \"4dcc939e-b548-4fb6-814a-30e2aaa8a94a\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-tmdj7" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.471343 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxln9\" (UniqueName: \"kubernetes.io/projected/967215e8-7a18-4507-96c1-4c79c7e1d51a-kube-api-access-xxln9\") pod \"heat-operator-controller-manager-5f64f6f8bb-d7c6w\" (UID: \"967215e8-7a18-4507-96c1-4c79c7e1d51a\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-d7c6w" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.471370 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4dcc939e-b548-4fb6-814a-30e2aaa8a94a-cert\") pod \"infra-operator-controller-manager-57548d458d-tmdj7\" (UID: \"4dcc939e-b548-4fb6-814a-30e2aaa8a94a\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-tmdj7" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.471409 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6phsr\" (UniqueName: \"kubernetes.io/projected/eb53d542-63e9-487d-9d06-237c4b2b9252-kube-api-access-6phsr\") pod \"ironic-operator-controller-manager-6c548fd776-7wmqg\" (UID: \"eb53d542-63e9-487d-9d06-237c4b2b9252\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7wmqg" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.471434 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgqgk\" (UniqueName: \"kubernetes.io/projected/94e7345b-b1ac-46e2-be25-9d64d3d33523-kube-api-access-qgqgk\") pod \"horizon-operator-controller-manager-68c6d99b8f-45whz\" (UID: \"94e7345b-b1ac-46e2-be25-9d64d3d33523\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-45whz" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.471479 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbjxp\" (UniqueName: \"kubernetes.io/projected/f778fd6c-4c8a-4067-b3f2-cb7d98a50bf7-kube-api-access-jbjxp\") pod \"glance-operator-controller-manager-5cd474b4ff-wcmv4\" (UID: \"f778fd6c-4c8a-4067-b3f2-cb7d98a50bf7\") " pod="openstack-operators/glance-operator-controller-manager-5cd474b4ff-wcmv4" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.523752 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbjxp\" (UniqueName: \"kubernetes.io/projected/f778fd6c-4c8a-4067-b3f2-cb7d98a50bf7-kube-api-access-jbjxp\") pod \"glance-operator-controller-manager-5cd474b4ff-wcmv4\" (UID: \"f778fd6c-4c8a-4067-b3f2-cb7d98a50bf7\") " pod="openstack-operators/glance-operator-controller-manager-5cd474b4ff-wcmv4" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.528844 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-7wmqg"] Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.533681 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-6546668bfd-j2vnt"] Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.534759 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxln9\" (UniqueName: \"kubernetes.io/projected/967215e8-7a18-4507-96c1-4c79c7e1d51a-kube-api-access-xxln9\") pod \"heat-operator-controller-manager-5f64f6f8bb-d7c6w\" (UID: \"967215e8-7a18-4507-96c1-4c79c7e1d51a\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-d7c6w" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.535266 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-j2vnt" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.538070 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgqgk\" (UniqueName: \"kubernetes.io/projected/94e7345b-b1ac-46e2-be25-9d64d3d33523-kube-api-access-qgqgk\") pod \"horizon-operator-controller-manager-68c6d99b8f-45whz\" (UID: \"94e7345b-b1ac-46e2-be25-9d64d3d33523\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-45whz" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.539843 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-sb29r" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.550095 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-hnzzf"] Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.551830 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-hnzzf" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.558397 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-546d4bdf48-pfphw"] Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.559925 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-pfphw" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.565987 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-j7fd8" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.573678 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6546668bfd-j2vnt"] Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.575669 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6phsr\" (UniqueName: \"kubernetes.io/projected/eb53d542-63e9-487d-9d06-237c4b2b9252-kube-api-access-6phsr\") pod \"ironic-operator-controller-manager-6c548fd776-7wmqg\" (UID: \"eb53d542-63e9-487d-9d06-237c4b2b9252\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7wmqg" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.575783 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9klvb\" (UniqueName: \"kubernetes.io/projected/4dcc939e-b548-4fb6-814a-30e2aaa8a94a-kube-api-access-9klvb\") pod \"infra-operator-controller-manager-57548d458d-tmdj7\" (UID: \"4dcc939e-b548-4fb6-814a-30e2aaa8a94a\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-tmdj7" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.575859 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4dcc939e-b548-4fb6-814a-30e2aaa8a94a-cert\") pod \"infra-operator-controller-manager-57548d458d-tmdj7\" (UID: \"4dcc939e-b548-4fb6-814a-30e2aaa8a94a\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-tmdj7" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.583815 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-5v74m" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.584005 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-546d4bdf48-pfphw"] Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.584189 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.585379 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-d7c6w" Dec 01 08:55:24 crc kubenswrapper[4873]: E1201 08:55:24.586085 4873 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 01 08:55:24 crc kubenswrapper[4873]: E1201 08:55:24.586153 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4dcc939e-b548-4fb6-814a-30e2aaa8a94a-cert podName:4dcc939e-b548-4fb6-814a-30e2aaa8a94a nodeName:}" failed. No retries permitted until 2025-12-01 08:55:25.086124533 +0000 UTC m=+900.988233072 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/4dcc939e-b548-4fb6-814a-30e2aaa8a94a-cert") pod "infra-operator-controller-manager-57548d458d-tmdj7" (UID: "4dcc939e-b548-4fb6-814a-30e2aaa8a94a") : secret "infra-operator-webhook-server-cert" not found Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.591144 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-rbr6s" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.595756 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6phsr\" (UniqueName: \"kubernetes.io/projected/eb53d542-63e9-487d-9d06-237c4b2b9252-kube-api-access-6phsr\") pod \"ironic-operator-controller-manager-6c548fd776-7wmqg\" (UID: \"eb53d542-63e9-487d-9d06-237c4b2b9252\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7wmqg" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.604497 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9klvb\" (UniqueName: \"kubernetes.io/projected/4dcc939e-b548-4fb6-814a-30e2aaa8a94a-kube-api-access-9klvb\") pod \"infra-operator-controller-manager-57548d458d-tmdj7\" (UID: \"4dcc939e-b548-4fb6-814a-30e2aaa8a94a\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-tmdj7" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.610953 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-hnzzf"] Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.617243 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-sjw48"] Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.618714 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-sjw48" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.624566 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-6qb7p" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.672346 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-x9kc4"] Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.689988 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-n79q9"] Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.691138 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-sjw48"] Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.691162 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vf8qt"] Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.692301 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-x9kc4" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.692744 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-n79q9" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.695697 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vf8qt" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.696368 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vlf2\" (UniqueName: \"kubernetes.io/projected/3eff7ce2-9ff7-413c-b472-9e114e7130ca-kube-api-access-6vlf2\") pod \"mariadb-operator-controller-manager-56bbcc9d85-hnzzf\" (UID: \"3eff7ce2-9ff7-413c-b472-9e114e7130ca\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-hnzzf" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.696399 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xc57g\" (UniqueName: \"kubernetes.io/projected/5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5-kube-api-access-xc57g\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4vf8qt\" (UID: \"5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vf8qt" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.696421 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4vf8qt\" (UID: \"5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vf8qt" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.696459 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hnhn\" (UniqueName: \"kubernetes.io/projected/7e3a2af8-3381-46e2-8c23-41aab8fd1a5e-kube-api-access-9hnhn\") pod \"octavia-operator-controller-manager-998648c74-n79q9\" (UID: \"7e3a2af8-3381-46e2-8c23-41aab8fd1a5e\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-n79q9" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.696483 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r57cm\" (UniqueName: \"kubernetes.io/projected/5ec1188a-c7f9-4cc8-8ea9-c1f1977041a5-kube-api-access-r57cm\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-sjw48\" (UID: \"5ec1188a-c7f9-4cc8-8ea9-c1f1977041a5\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-sjw48" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.696520 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7tkm\" (UniqueName: \"kubernetes.io/projected/b45909ce-2ea4-4cf2-9351-c8839e44d734-kube-api-access-k7tkm\") pod \"manila-operator-controller-manager-6546668bfd-j2vnt\" (UID: \"b45909ce-2ea4-4cf2-9351-c8839e44d734\") " pod="openstack-operators/manila-operator-controller-manager-6546668bfd-j2vnt" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.696569 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9k7x\" (UniqueName: \"kubernetes.io/projected/450a3417-0074-4223-b1e3-aa1b854320fe-kube-api-access-j9k7x\") pod \"nova-operator-controller-manager-697bc559fc-x9kc4\" (UID: \"450a3417-0074-4223-b1e3-aa1b854320fe\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-x9kc4" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.696589 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xtkm\" (UniqueName: \"kubernetes.io/projected/035095bd-1b65-4895-95a1-59feee524920-kube-api-access-8xtkm\") pod \"keystone-operator-controller-manager-546d4bdf48-pfphw\" (UID: \"035095bd-1b65-4895-95a1-59feee524920\") " pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-pfphw" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.697890 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-2mcxc" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.698266 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.699512 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-55dm7" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.701561 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-x9kc4"] Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.705110 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-xbxs7" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.705611 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-mmpkl" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.710198 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-45whz" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.718628 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vf8qt"] Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.728073 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-n79q9"] Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.735039 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7wmqg" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.755836 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-v7bvz"] Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.788894 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-v7bvz" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.822764 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-m54vn" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.825884 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-v7bvz"] Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.858504 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-x4w4m" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.849484 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vlf2\" (UniqueName: \"kubernetes.io/projected/3eff7ce2-9ff7-413c-b472-9e114e7130ca-kube-api-access-6vlf2\") pod \"mariadb-operator-controller-manager-56bbcc9d85-hnzzf\" (UID: \"3eff7ce2-9ff7-413c-b472-9e114e7130ca\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-hnzzf" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.967566 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5cd474b4ff-wcmv4" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.867395 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xc57g\" (UniqueName: \"kubernetes.io/projected/5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5-kube-api-access-xc57g\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4vf8qt\" (UID: \"5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vf8qt" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.995983 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4vf8qt\" (UID: \"5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vf8qt" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.996310 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hnhn\" (UniqueName: \"kubernetes.io/projected/7e3a2af8-3381-46e2-8c23-41aab8fd1a5e-kube-api-access-9hnhn\") pod \"octavia-operator-controller-manager-998648c74-n79q9\" (UID: \"7e3a2af8-3381-46e2-8c23-41aab8fd1a5e\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-n79q9" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.996361 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r57cm\" (UniqueName: \"kubernetes.io/projected/5ec1188a-c7f9-4cc8-8ea9-c1f1977041a5-kube-api-access-r57cm\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-sjw48\" (UID: \"5ec1188a-c7f9-4cc8-8ea9-c1f1977041a5\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-sjw48" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.996434 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7tkm\" (UniqueName: \"kubernetes.io/projected/b45909ce-2ea4-4cf2-9351-c8839e44d734-kube-api-access-k7tkm\") pod \"manila-operator-controller-manager-6546668bfd-j2vnt\" (UID: \"b45909ce-2ea4-4cf2-9351-c8839e44d734\") " pod="openstack-operators/manila-operator-controller-manager-6546668bfd-j2vnt" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.996520 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9k7x\" (UniqueName: \"kubernetes.io/projected/450a3417-0074-4223-b1e3-aa1b854320fe-kube-api-access-j9k7x\") pod \"nova-operator-controller-manager-697bc559fc-x9kc4\" (UID: \"450a3417-0074-4223-b1e3-aa1b854320fe\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-x9kc4" Dec 01 08:55:24 crc kubenswrapper[4873]: I1201 08:55:24.996554 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xtkm\" (UniqueName: \"kubernetes.io/projected/035095bd-1b65-4895-95a1-59feee524920-kube-api-access-8xtkm\") pod \"keystone-operator-controller-manager-546d4bdf48-pfphw\" (UID: \"035095bd-1b65-4895-95a1-59feee524920\") " pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-pfphw" Dec 01 08:55:24 crc kubenswrapper[4873]: E1201 08:55:24.997175 4873 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 08:55:24 crc kubenswrapper[4873]: E1201 08:55:24.997237 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5-cert podName:5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5 nodeName:}" failed. No retries permitted until 2025-12-01 08:55:25.497220489 +0000 UTC m=+901.399329028 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4vf8qt" (UID: "5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.016450 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vlf2\" (UniqueName: \"kubernetes.io/projected/3eff7ce2-9ff7-413c-b472-9e114e7130ca-kube-api-access-6vlf2\") pod \"mariadb-operator-controller-manager-56bbcc9d85-hnzzf\" (UID: \"3eff7ce2-9ff7-413c-b472-9e114e7130ca\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-hnzzf" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.036450 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-49bz9"] Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.039206 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-49bz9" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.062877 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xc57g\" (UniqueName: \"kubernetes.io/projected/5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5-kube-api-access-xc57g\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4vf8qt\" (UID: \"5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vf8qt" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.246935 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9k7x\" (UniqueName: \"kubernetes.io/projected/450a3417-0074-4223-b1e3-aa1b854320fe-kube-api-access-j9k7x\") pod \"nova-operator-controller-manager-697bc559fc-x9kc4\" (UID: \"450a3417-0074-4223-b1e3-aa1b854320fe\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-x9kc4" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.249655 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7tkm\" (UniqueName: \"kubernetes.io/projected/b45909ce-2ea4-4cf2-9351-c8839e44d734-kube-api-access-k7tkm\") pod \"manila-operator-controller-manager-6546668bfd-j2vnt\" (UID: \"b45909ce-2ea4-4cf2-9351-c8839e44d734\") " pod="openstack-operators/manila-operator-controller-manager-6546668bfd-j2vnt" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.251692 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-hnzzf" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.257091 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bl8tk\" (UniqueName: \"kubernetes.io/projected/879e02ea-306e-4e7b-9012-cb1cb6bdee00-kube-api-access-bl8tk\") pod \"ovn-operator-controller-manager-b6456fdb6-49bz9\" (UID: \"879e02ea-306e-4e7b-9012-cb1cb6bdee00\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-49bz9" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.257191 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4dcc939e-b548-4fb6-814a-30e2aaa8a94a-cert\") pod \"infra-operator-controller-manager-57548d458d-tmdj7\" (UID: \"4dcc939e-b548-4fb6-814a-30e2aaa8a94a\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-tmdj7" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.257228 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhkzv\" (UniqueName: \"kubernetes.io/projected/3e0ad069-d971-4fd0-93df-b6a8794afe00-kube-api-access-hhkzv\") pod \"placement-operator-controller-manager-78f8948974-v7bvz\" (UID: \"3e0ad069-d971-4fd0-93df-b6a8794afe00\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-v7bvz" Dec 01 08:55:25 crc kubenswrapper[4873]: E1201 08:55:25.257469 4873 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 01 08:55:25 crc kubenswrapper[4873]: E1201 08:55:25.257526 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4dcc939e-b548-4fb6-814a-30e2aaa8a94a-cert podName:4dcc939e-b548-4fb6-814a-30e2aaa8a94a nodeName:}" failed. No retries permitted until 2025-12-01 08:55:26.257506804 +0000 UTC m=+902.159615343 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/4dcc939e-b548-4fb6-814a-30e2aaa8a94a-cert") pod "infra-operator-controller-manager-57548d458d-tmdj7" (UID: "4dcc939e-b548-4fb6-814a-30e2aaa8a94a") : secret "infra-operator-webhook-server-cert" not found Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.261426 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xtkm\" (UniqueName: \"kubernetes.io/projected/035095bd-1b65-4895-95a1-59feee524920-kube-api-access-8xtkm\") pod \"keystone-operator-controller-manager-546d4bdf48-pfphw\" (UID: \"035095bd-1b65-4895-95a1-59feee524920\") " pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-pfphw" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.262856 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hnhn\" (UniqueName: \"kubernetes.io/projected/7e3a2af8-3381-46e2-8c23-41aab8fd1a5e-kube-api-access-9hnhn\") pod \"octavia-operator-controller-manager-998648c74-n79q9\" (UID: \"7e3a2af8-3381-46e2-8c23-41aab8fd1a5e\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-n79q9" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.262866 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-j5gbg" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.268165 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r57cm\" (UniqueName: \"kubernetes.io/projected/5ec1188a-c7f9-4cc8-8ea9-c1f1977041a5-kube-api-access-r57cm\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-sjw48\" (UID: \"5ec1188a-c7f9-4cc8-8ea9-c1f1977041a5\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-sjw48" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.280126 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6dxcw"] Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.281590 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6dxcw" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.308990 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-49bz9"] Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.309096 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-bhrpn"] Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.310609 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6dxcw"] Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.310741 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-bhrpn" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.328409 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-bhrpn"] Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.336838 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-x9kc4" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.337259 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-tt55p" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.337539 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-kwzsm" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.337802 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-sjw48" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.359674 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bl8tk\" (UniqueName: \"kubernetes.io/projected/879e02ea-306e-4e7b-9012-cb1cb6bdee00-kube-api-access-bl8tk\") pod \"ovn-operator-controller-manager-b6456fdb6-49bz9\" (UID: \"879e02ea-306e-4e7b-9012-cb1cb6bdee00\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-49bz9" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.359760 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmb2k\" (UniqueName: \"kubernetes.io/projected/225323c8-ed2f-4573-920d-f43f3e4561de-kube-api-access-rmb2k\") pod \"telemetry-operator-controller-manager-76cc84c6bb-bhrpn\" (UID: \"225323c8-ed2f-4573-920d-f43f3e4561de\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-bhrpn" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.359858 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhkzv\" (UniqueName: \"kubernetes.io/projected/3e0ad069-d971-4fd0-93df-b6a8794afe00-kube-api-access-hhkzv\") pod \"placement-operator-controller-manager-78f8948974-v7bvz\" (UID: \"3e0ad069-d971-4fd0-93df-b6a8794afe00\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-v7bvz" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.359884 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrqtf\" (UniqueName: \"kubernetes.io/projected/f7a7bcca-9403-4e2d-bc28-53d4eb5ae252-kube-api-access-nrqtf\") pod \"swift-operator-controller-manager-5f8c65bbfc-6dxcw\" (UID: \"f7a7bcca-9403-4e2d-bc28-53d4eb5ae252\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6dxcw" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.361166 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-n79q9" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.536457 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-mvc26"] Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.538690 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-mvc26" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.558711 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-dx4jm"] Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.560296 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-dx4jm" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.577002 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-dx4jm"] Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.588259 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-mvc26"] Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.603723 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-j2vnt" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.604370 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrqtf\" (UniqueName: \"kubernetes.io/projected/f7a7bcca-9403-4e2d-bc28-53d4eb5ae252-kube-api-access-nrqtf\") pod \"swift-operator-controller-manager-5f8c65bbfc-6dxcw\" (UID: \"f7a7bcca-9403-4e2d-bc28-53d4eb5ae252\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6dxcw" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.604433 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4vf8qt\" (UID: \"5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vf8qt" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.604845 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhkzv\" (UniqueName: \"kubernetes.io/projected/3e0ad069-d971-4fd0-93df-b6a8794afe00-kube-api-access-hhkzv\") pod \"placement-operator-controller-manager-78f8948974-v7bvz\" (UID: \"3e0ad069-d971-4fd0-93df-b6a8794afe00\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-v7bvz" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.604860 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmb2k\" (UniqueName: \"kubernetes.io/projected/225323c8-ed2f-4573-920d-f43f3e4561de-kube-api-access-rmb2k\") pod \"telemetry-operator-controller-manager-76cc84c6bb-bhrpn\" (UID: \"225323c8-ed2f-4573-920d-f43f3e4561de\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-bhrpn" Dec 01 08:55:25 crc kubenswrapper[4873]: E1201 08:55:25.605618 4873 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.605612 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7f4d6cf897-fddkc"] Dec 01 08:55:25 crc kubenswrapper[4873]: E1201 08:55:25.605667 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5-cert podName:5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5 nodeName:}" failed. No retries permitted until 2025-12-01 08:55:26.605652605 +0000 UTC m=+902.507761144 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4vf8qt" (UID: "5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.606954 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-pfphw" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.607994 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-v7bvz" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.603740 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bl8tk\" (UniqueName: \"kubernetes.io/projected/879e02ea-306e-4e7b-9012-cb1cb6bdee00-kube-api-access-bl8tk\") pod \"ovn-operator-controller-manager-b6456fdb6-49bz9\" (UID: \"879e02ea-306e-4e7b-9012-cb1cb6bdee00\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-49bz9" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.617263 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7f4d6cf897-fddkc" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.641940 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-x2hpj" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.655957 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-d2gmj" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.656332 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.656433 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.656562 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-sk567" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.658643 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-49bz9" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.680666 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrqtf\" (UniqueName: \"kubernetes.io/projected/f7a7bcca-9403-4e2d-bc28-53d4eb5ae252-kube-api-access-nrqtf\") pod \"swift-operator-controller-manager-5f8c65bbfc-6dxcw\" (UID: \"f7a7bcca-9403-4e2d-bc28-53d4eb5ae252\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6dxcw" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.706754 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/223a64ee-92cb-4b29-91bf-ffa7ed7d64ce-webhook-certs\") pod \"openstack-operator-controller-manager-7f4d6cf897-fddkc\" (UID: \"223a64ee-92cb-4b29-91bf-ffa7ed7d64ce\") " pod="openstack-operators/openstack-operator-controller-manager-7f4d6cf897-fddkc" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.707254 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/223a64ee-92cb-4b29-91bf-ffa7ed7d64ce-metrics-certs\") pod \"openstack-operator-controller-manager-7f4d6cf897-fddkc\" (UID: \"223a64ee-92cb-4b29-91bf-ffa7ed7d64ce\") " pod="openstack-operators/openstack-operator-controller-manager-7f4d6cf897-fddkc" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.707278 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89mn9\" (UniqueName: \"kubernetes.io/projected/223a64ee-92cb-4b29-91bf-ffa7ed7d64ce-kube-api-access-89mn9\") pod \"openstack-operator-controller-manager-7f4d6cf897-fddkc\" (UID: \"223a64ee-92cb-4b29-91bf-ffa7ed7d64ce\") " pod="openstack-operators/openstack-operator-controller-manager-7f4d6cf897-fddkc" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.707367 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-994kd\" (UniqueName: \"kubernetes.io/projected/1cd3aa85-6333-4cfd-aacf-e51169ef4b42-kube-api-access-994kd\") pod \"watcher-operator-controller-manager-769dc69bc-dx4jm\" (UID: \"1cd3aa85-6333-4cfd-aacf-e51169ef4b42\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-dx4jm" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.707403 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hh8b2\" (UniqueName: \"kubernetes.io/projected/34e3726b-c9f5-4a86-9e87-e179f8dae739-kube-api-access-hh8b2\") pod \"test-operator-controller-manager-5854674fcc-mvc26\" (UID: \"34e3726b-c9f5-4a86-9e87-e179f8dae739\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-mvc26" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.732274 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmb2k\" (UniqueName: \"kubernetes.io/projected/225323c8-ed2f-4573-920d-f43f3e4561de-kube-api-access-rmb2k\") pod \"telemetry-operator-controller-manager-76cc84c6bb-bhrpn\" (UID: \"225323c8-ed2f-4573-920d-f43f3e4561de\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-bhrpn" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.732356 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7f4d6cf897-fddkc"] Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.739889 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cb9qn"] Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.744263 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cb9qn" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.752174 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-6jr7p" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.771105 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cb9qn"] Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.809185 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/223a64ee-92cb-4b29-91bf-ffa7ed7d64ce-metrics-certs\") pod \"openstack-operator-controller-manager-7f4d6cf897-fddkc\" (UID: \"223a64ee-92cb-4b29-91bf-ffa7ed7d64ce\") " pod="openstack-operators/openstack-operator-controller-manager-7f4d6cf897-fddkc" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.809235 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89mn9\" (UniqueName: \"kubernetes.io/projected/223a64ee-92cb-4b29-91bf-ffa7ed7d64ce-kube-api-access-89mn9\") pod \"openstack-operator-controller-manager-7f4d6cf897-fddkc\" (UID: \"223a64ee-92cb-4b29-91bf-ffa7ed7d64ce\") " pod="openstack-operators/openstack-operator-controller-manager-7f4d6cf897-fddkc" Dec 01 08:55:25 crc kubenswrapper[4873]: E1201 08:55:25.809295 4873 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.809350 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-994kd\" (UniqueName: \"kubernetes.io/projected/1cd3aa85-6333-4cfd-aacf-e51169ef4b42-kube-api-access-994kd\") pod \"watcher-operator-controller-manager-769dc69bc-dx4jm\" (UID: \"1cd3aa85-6333-4cfd-aacf-e51169ef4b42\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-dx4jm" Dec 01 08:55:25 crc kubenswrapper[4873]: E1201 08:55:25.809393 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/223a64ee-92cb-4b29-91bf-ffa7ed7d64ce-metrics-certs podName:223a64ee-92cb-4b29-91bf-ffa7ed7d64ce nodeName:}" failed. No retries permitted until 2025-12-01 08:55:26.309368812 +0000 UTC m=+902.211477541 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/223a64ee-92cb-4b29-91bf-ffa7ed7d64ce-metrics-certs") pod "openstack-operator-controller-manager-7f4d6cf897-fddkc" (UID: "223a64ee-92cb-4b29-91bf-ffa7ed7d64ce") : secret "metrics-server-cert" not found Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.809427 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hh8b2\" (UniqueName: \"kubernetes.io/projected/34e3726b-c9f5-4a86-9e87-e179f8dae739-kube-api-access-hh8b2\") pod \"test-operator-controller-manager-5854674fcc-mvc26\" (UID: \"34e3726b-c9f5-4a86-9e87-e179f8dae739\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-mvc26" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.809519 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/223a64ee-92cb-4b29-91bf-ffa7ed7d64ce-webhook-certs\") pod \"openstack-operator-controller-manager-7f4d6cf897-fddkc\" (UID: \"223a64ee-92cb-4b29-91bf-ffa7ed7d64ce\") " pod="openstack-operators/openstack-operator-controller-manager-7f4d6cf897-fddkc" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.809555 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klhwp\" (UniqueName: \"kubernetes.io/projected/664e20ef-b15c-4903-b72d-e18c7077e888-kube-api-access-klhwp\") pod \"rabbitmq-cluster-operator-manager-668c99d594-cb9qn\" (UID: \"664e20ef-b15c-4903-b72d-e18c7077e888\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cb9qn" Dec 01 08:55:25 crc kubenswrapper[4873]: E1201 08:55:25.809871 4873 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 01 08:55:25 crc kubenswrapper[4873]: E1201 08:55:25.809924 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/223a64ee-92cb-4b29-91bf-ffa7ed7d64ce-webhook-certs podName:223a64ee-92cb-4b29-91bf-ffa7ed7d64ce nodeName:}" failed. No retries permitted until 2025-12-01 08:55:26.309906996 +0000 UTC m=+902.212015535 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/223a64ee-92cb-4b29-91bf-ffa7ed7d64ce-webhook-certs") pod "openstack-operator-controller-manager-7f4d6cf897-fddkc" (UID: "223a64ee-92cb-4b29-91bf-ffa7ed7d64ce") : secret "webhook-server-cert" not found Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.811613 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6dxcw" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.867701 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hh8b2\" (UniqueName: \"kubernetes.io/projected/34e3726b-c9f5-4a86-9e87-e179f8dae739-kube-api-access-hh8b2\") pod \"test-operator-controller-manager-5854674fcc-mvc26\" (UID: \"34e3726b-c9f5-4a86-9e87-e179f8dae739\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-mvc26" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.872288 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-bhrpn" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.878316 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-994kd\" (UniqueName: \"kubernetes.io/projected/1cd3aa85-6333-4cfd-aacf-e51169ef4b42-kube-api-access-994kd\") pod \"watcher-operator-controller-manager-769dc69bc-dx4jm\" (UID: \"1cd3aa85-6333-4cfd-aacf-e51169ef4b42\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-dx4jm" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.879041 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89mn9\" (UniqueName: \"kubernetes.io/projected/223a64ee-92cb-4b29-91bf-ffa7ed7d64ce-kube-api-access-89mn9\") pod \"openstack-operator-controller-manager-7f4d6cf897-fddkc\" (UID: \"223a64ee-92cb-4b29-91bf-ffa7ed7d64ce\") " pod="openstack-operators/openstack-operator-controller-manager-7f4d6cf897-fddkc" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.902884 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-mvc26" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.917093 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klhwp\" (UniqueName: \"kubernetes.io/projected/664e20ef-b15c-4903-b72d-e18c7077e888-kube-api-access-klhwp\") pod \"rabbitmq-cluster-operator-manager-668c99d594-cb9qn\" (UID: \"664e20ef-b15c-4903-b72d-e18c7077e888\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cb9qn" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.931124 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-dx4jm" Dec 01 08:55:25 crc kubenswrapper[4873]: I1201 08:55:25.933946 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klhwp\" (UniqueName: \"kubernetes.io/projected/664e20ef-b15c-4903-b72d-e18c7077e888-kube-api-access-klhwp\") pod \"rabbitmq-cluster-operator-manager-668c99d594-cb9qn\" (UID: \"664e20ef-b15c-4903-b72d-e18c7077e888\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cb9qn" Dec 01 08:55:26 crc kubenswrapper[4873]: I1201 08:55:26.001778 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cb9qn" Dec 01 08:55:26 crc kubenswrapper[4873]: I1201 08:55:26.071768 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-5rq8z"] Dec 01 08:55:26 crc kubenswrapper[4873]: I1201 08:55:26.295670 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-gqm5w"] Dec 01 08:55:26 crc kubenswrapper[4873]: I1201 08:55:26.347714 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/223a64ee-92cb-4b29-91bf-ffa7ed7d64ce-metrics-certs\") pod \"openstack-operator-controller-manager-7f4d6cf897-fddkc\" (UID: \"223a64ee-92cb-4b29-91bf-ffa7ed7d64ce\") " pod="openstack-operators/openstack-operator-controller-manager-7f4d6cf897-fddkc" Dec 01 08:55:26 crc kubenswrapper[4873]: I1201 08:55:26.347929 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4dcc939e-b548-4fb6-814a-30e2aaa8a94a-cert\") pod \"infra-operator-controller-manager-57548d458d-tmdj7\" (UID: \"4dcc939e-b548-4fb6-814a-30e2aaa8a94a\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-tmdj7" Dec 01 08:55:26 crc kubenswrapper[4873]: I1201 08:55:26.348108 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/223a64ee-92cb-4b29-91bf-ffa7ed7d64ce-webhook-certs\") pod \"openstack-operator-controller-manager-7f4d6cf897-fddkc\" (UID: \"223a64ee-92cb-4b29-91bf-ffa7ed7d64ce\") " pod="openstack-operators/openstack-operator-controller-manager-7f4d6cf897-fddkc" Dec 01 08:55:26 crc kubenswrapper[4873]: E1201 08:55:26.348342 4873 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 01 08:55:26 crc kubenswrapper[4873]: E1201 08:55:26.348487 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/223a64ee-92cb-4b29-91bf-ffa7ed7d64ce-webhook-certs podName:223a64ee-92cb-4b29-91bf-ffa7ed7d64ce nodeName:}" failed. No retries permitted until 2025-12-01 08:55:27.348466613 +0000 UTC m=+903.250575152 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/223a64ee-92cb-4b29-91bf-ffa7ed7d64ce-webhook-certs") pod "openstack-operator-controller-manager-7f4d6cf897-fddkc" (UID: "223a64ee-92cb-4b29-91bf-ffa7ed7d64ce") : secret "webhook-server-cert" not found Dec 01 08:55:26 crc kubenswrapper[4873]: E1201 08:55:26.348677 4873 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 01 08:55:26 crc kubenswrapper[4873]: E1201 08:55:26.348776 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/223a64ee-92cb-4b29-91bf-ffa7ed7d64ce-metrics-certs podName:223a64ee-92cb-4b29-91bf-ffa7ed7d64ce nodeName:}" failed. No retries permitted until 2025-12-01 08:55:27.34875279 +0000 UTC m=+903.250861329 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/223a64ee-92cb-4b29-91bf-ffa7ed7d64ce-metrics-certs") pod "openstack-operator-controller-manager-7f4d6cf897-fddkc" (UID: "223a64ee-92cb-4b29-91bf-ffa7ed7d64ce") : secret "metrics-server-cert" not found Dec 01 08:55:26 crc kubenswrapper[4873]: E1201 08:55:26.348832 4873 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 01 08:55:26 crc kubenswrapper[4873]: E1201 08:55:26.348862 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4dcc939e-b548-4fb6-814a-30e2aaa8a94a-cert podName:4dcc939e-b548-4fb6-814a-30e2aaa8a94a nodeName:}" failed. No retries permitted until 2025-12-01 08:55:28.348855093 +0000 UTC m=+904.250963632 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/4dcc939e-b548-4fb6-814a-30e2aaa8a94a-cert") pod "infra-operator-controller-manager-57548d458d-tmdj7" (UID: "4dcc939e-b548-4fb6-814a-30e2aaa8a94a") : secret "infra-operator-webhook-server-cert" not found Dec 01 08:55:26 crc kubenswrapper[4873]: W1201 08:55:26.415256 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod671ad3cc_9716_42de_b5ae_fb69847d3bd7.slice/crio-91f2d1eaf49477ffa406d5f43fe0fcbc48338e3ae0ffd60205b040649fd54af3 WatchSource:0}: Error finding container 91f2d1eaf49477ffa406d5f43fe0fcbc48338e3ae0ffd60205b040649fd54af3: Status 404 returned error can't find the container with id 91f2d1eaf49477ffa406d5f43fe0fcbc48338e3ae0ffd60205b040649fd54af3 Dec 01 08:55:26 crc kubenswrapper[4873]: I1201 08:55:26.635140 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4vf8qt\" (UID: \"5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vf8qt" Dec 01 08:55:26 crc kubenswrapper[4873]: E1201 08:55:26.635392 4873 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 08:55:26 crc kubenswrapper[4873]: E1201 08:55:26.635536 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5-cert podName:5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5 nodeName:}" failed. No retries permitted until 2025-12-01 08:55:28.635500764 +0000 UTC m=+904.537609303 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4vf8qt" (UID: "5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 08:55:26 crc kubenswrapper[4873]: I1201 08:55:26.753883 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-6r8hb"] Dec 01 08:55:26 crc kubenswrapper[4873]: I1201 08:55:26.852257 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-gqm5w" event={"ID":"671ad3cc-9716-42de-b5ae-fb69847d3bd7","Type":"ContainerStarted","Data":"91f2d1eaf49477ffa406d5f43fe0fcbc48338e3ae0ffd60205b040649fd54af3"} Dec 01 08:55:26 crc kubenswrapper[4873]: I1201 08:55:26.854686 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5rq8z" event={"ID":"51253ba6-e0b5-44ac-8c18-be17b4d13024","Type":"ContainerStarted","Data":"9bd6cc1f4525aedcd462104b88528d6266fbfc1a4ab20d839232072f63a5332d"} Dec 01 08:55:27 crc kubenswrapper[4873]: I1201 08:55:27.351568 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/223a64ee-92cb-4b29-91bf-ffa7ed7d64ce-webhook-certs\") pod \"openstack-operator-controller-manager-7f4d6cf897-fddkc\" (UID: \"223a64ee-92cb-4b29-91bf-ffa7ed7d64ce\") " pod="openstack-operators/openstack-operator-controller-manager-7f4d6cf897-fddkc" Dec 01 08:55:27 crc kubenswrapper[4873]: I1201 08:55:27.351645 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/223a64ee-92cb-4b29-91bf-ffa7ed7d64ce-metrics-certs\") pod \"openstack-operator-controller-manager-7f4d6cf897-fddkc\" (UID: \"223a64ee-92cb-4b29-91bf-ffa7ed7d64ce\") " pod="openstack-operators/openstack-operator-controller-manager-7f4d6cf897-fddkc" Dec 01 08:55:27 crc kubenswrapper[4873]: E1201 08:55:27.351808 4873 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 01 08:55:27 crc kubenswrapper[4873]: E1201 08:55:27.351889 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/223a64ee-92cb-4b29-91bf-ffa7ed7d64ce-metrics-certs podName:223a64ee-92cb-4b29-91bf-ffa7ed7d64ce nodeName:}" failed. No retries permitted until 2025-12-01 08:55:29.351870804 +0000 UTC m=+905.253979343 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/223a64ee-92cb-4b29-91bf-ffa7ed7d64ce-metrics-certs") pod "openstack-operator-controller-manager-7f4d6cf897-fddkc" (UID: "223a64ee-92cb-4b29-91bf-ffa7ed7d64ce") : secret "metrics-server-cert" not found Dec 01 08:55:27 crc kubenswrapper[4873]: E1201 08:55:27.352304 4873 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 01 08:55:27 crc kubenswrapper[4873]: E1201 08:55:27.352401 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/223a64ee-92cb-4b29-91bf-ffa7ed7d64ce-webhook-certs podName:223a64ee-92cb-4b29-91bf-ffa7ed7d64ce nodeName:}" failed. No retries permitted until 2025-12-01 08:55:29.352377107 +0000 UTC m=+905.254485636 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/223a64ee-92cb-4b29-91bf-ffa7ed7d64ce-webhook-certs") pod "openstack-operator-controller-manager-7f4d6cf897-fddkc" (UID: "223a64ee-92cb-4b29-91bf-ffa7ed7d64ce") : secret "webhook-server-cert" not found Dec 01 08:55:27 crc kubenswrapper[4873]: I1201 08:55:27.861058 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-7wmqg"] Dec 01 08:55:27 crc kubenswrapper[4873]: I1201 08:55:27.878201 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-6r8hb" event={"ID":"723f446d-21bb-432a-b2af-93219593819d","Type":"ContainerStarted","Data":"05740462cff923ce6b3df40bc65933aae97c2adf20b382b917233e0d1cd6a312"} Dec 01 08:55:27 crc kubenswrapper[4873]: I1201 08:55:27.885778 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5cd474b4ff-wcmv4"] Dec 01 08:55:27 crc kubenswrapper[4873]: I1201 08:55:27.958545 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-x9kc4"] Dec 01 08:55:28 crc kubenswrapper[4873]: I1201 08:55:28.030734 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-d7c6w"] Dec 01 08:55:28 crc kubenswrapper[4873]: I1201 08:55:28.055921 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-sjw48"] Dec 01 08:55:28 crc kubenswrapper[4873]: I1201 08:55:28.241482 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6546668bfd-j2vnt"] Dec 01 08:55:28 crc kubenswrapper[4873]: I1201 08:55:28.267466 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-45whz"] Dec 01 08:55:28 crc kubenswrapper[4873]: I1201 08:55:28.280645 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-n79q9"] Dec 01 08:55:28 crc kubenswrapper[4873]: I1201 08:55:28.308991 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-hnzzf"] Dec 01 08:55:28 crc kubenswrapper[4873]: W1201 08:55:28.329978 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7e3a2af8_3381_46e2_8c23_41aab8fd1a5e.slice/crio-f37561c71cc4e294cde8f00a1cb458b98f53b6a61a36b9bd46f627d8ea8acb48 WatchSource:0}: Error finding container f37561c71cc4e294cde8f00a1cb458b98f53b6a61a36b9bd46f627d8ea8acb48: Status 404 returned error can't find the container with id f37561c71cc4e294cde8f00a1cb458b98f53b6a61a36b9bd46f627d8ea8acb48 Dec 01 08:55:28 crc kubenswrapper[4873]: I1201 08:55:28.397387 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-bhrpn"] Dec 01 08:55:28 crc kubenswrapper[4873]: I1201 08:55:28.410908 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4dcc939e-b548-4fb6-814a-30e2aaa8a94a-cert\") pod \"infra-operator-controller-manager-57548d458d-tmdj7\" (UID: \"4dcc939e-b548-4fb6-814a-30e2aaa8a94a\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-tmdj7" Dec 01 08:55:28 crc kubenswrapper[4873]: E1201 08:55:28.410900 4873 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 01 08:55:28 crc kubenswrapper[4873]: E1201 08:55:28.411145 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4dcc939e-b548-4fb6-814a-30e2aaa8a94a-cert podName:4dcc939e-b548-4fb6-814a-30e2aaa8a94a nodeName:}" failed. No retries permitted until 2025-12-01 08:55:32.411095643 +0000 UTC m=+908.313204172 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/4dcc939e-b548-4fb6-814a-30e2aaa8a94a-cert") pod "infra-operator-controller-manager-57548d458d-tmdj7" (UID: "4dcc939e-b548-4fb6-814a-30e2aaa8a94a") : secret "infra-operator-webhook-server-cert" not found Dec 01 08:55:28 crc kubenswrapper[4873]: I1201 08:55:28.421046 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-mvc26"] Dec 01 08:55:28 crc kubenswrapper[4873]: I1201 08:55:28.549941 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-546d4bdf48-pfphw"] Dec 01 08:55:28 crc kubenswrapper[4873]: I1201 08:55:28.575756 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-dx4jm"] Dec 01 08:55:28 crc kubenswrapper[4873]: I1201 08:55:28.610321 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6dxcw"] Dec 01 08:55:28 crc kubenswrapper[4873]: I1201 08:55:28.617998 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cb9qn"] Dec 01 08:55:28 crc kubenswrapper[4873]: I1201 08:55:28.661470 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-v7bvz"] Dec 01 08:55:28 crc kubenswrapper[4873]: E1201 08:55:28.670761 4873 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-klhwp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-cb9qn_openstack-operators(664e20ef-b15c-4903-b72d-e18c7077e888): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 08:55:28 crc kubenswrapper[4873]: E1201 08:55:28.673843 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cb9qn" podUID="664e20ef-b15c-4903-b72d-e18c7077e888" Dec 01 08:55:28 crc kubenswrapper[4873]: E1201 08:55:28.715144 4873 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nrqtf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-6dxcw_openstack-operators(f7a7bcca-9403-4e2d-bc28-53d4eb5ae252): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 08:55:28 crc kubenswrapper[4873]: I1201 08:55:28.733151 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-49bz9"] Dec 01 08:55:28 crc kubenswrapper[4873]: I1201 08:55:28.746391 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4vf8qt\" (UID: \"5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vf8qt" Dec 01 08:55:28 crc kubenswrapper[4873]: E1201 08:55:28.746698 4873 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 08:55:28 crc kubenswrapper[4873]: E1201 08:55:28.746781 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5-cert podName:5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5 nodeName:}" failed. No retries permitted until 2025-12-01 08:55:32.746754763 +0000 UTC m=+908.648863302 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4vf8qt" (UID: "5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 08:55:28 crc kubenswrapper[4873]: E1201 08:55:28.789512 4873 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hhkzv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-v7bvz_openstack-operators(3e0ad069-d971-4fd0-93df-b6a8794afe00): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 08:55:28 crc kubenswrapper[4873]: E1201 08:55:28.811405 4873 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hhkzv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-v7bvz_openstack-operators(3e0ad069-d971-4fd0-93df-b6a8794afe00): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 08:55:28 crc kubenswrapper[4873]: E1201 08:55:28.812541 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-v7bvz" podUID="3e0ad069-d971-4fd0-93df-b6a8794afe00" Dec 01 08:55:28 crc kubenswrapper[4873]: E1201 08:55:28.813058 4873 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bl8tk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-49bz9_openstack-operators(879e02ea-306e-4e7b-9012-cb1cb6bdee00): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 08:55:28 crc kubenswrapper[4873]: E1201 08:55:28.827605 4873 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bl8tk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-49bz9_openstack-operators(879e02ea-306e-4e7b-9012-cb1cb6bdee00): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 08:55:28 crc kubenswrapper[4873]: E1201 08:55:28.829096 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-49bz9" podUID="879e02ea-306e-4e7b-9012-cb1cb6bdee00" Dec 01 08:55:28 crc kubenswrapper[4873]: I1201 08:55:28.945921 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-hnzzf" event={"ID":"3eff7ce2-9ff7-413c-b472-9e114e7130ca","Type":"ContainerStarted","Data":"7cf577dfe6201027660eea9a8af94beefa5efd2c7d6e0c2889798fed8360db12"} Dec 01 08:55:29 crc kubenswrapper[4873]: I1201 08:55:29.003395 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-mvc26" event={"ID":"34e3726b-c9f5-4a86-9e87-e179f8dae739","Type":"ContainerStarted","Data":"030feaefe8b296d88e40117f20e73770823061b7b1d42e58085b26ed04421435"} Dec 01 08:55:29 crc kubenswrapper[4873]: I1201 08:55:29.045342 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5cd474b4ff-wcmv4" event={"ID":"f778fd6c-4c8a-4067-b3f2-cb7d98a50bf7","Type":"ContainerStarted","Data":"f6326f9245f13b2085166513f4944b20d27ae06a85d63b5c8bd8da9f0d43ecf0"} Dec 01 08:55:29 crc kubenswrapper[4873]: I1201 08:55:29.048746 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-bhrpn" event={"ID":"225323c8-ed2f-4573-920d-f43f3e4561de","Type":"ContainerStarted","Data":"9bab0970104e6a6386d4dfebd2e05095d36f619ff8bd3fb21f553f069e8652d1"} Dec 01 08:55:29 crc kubenswrapper[4873]: I1201 08:55:29.049779 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cb9qn" event={"ID":"664e20ef-b15c-4903-b72d-e18c7077e888","Type":"ContainerStarted","Data":"a1cbab852e3478b29793ff349b3e4d503e057293a858a258ee250e3f1b92c636"} Dec 01 08:55:29 crc kubenswrapper[4873]: E1201 08:55:29.050904 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cb9qn" podUID="664e20ef-b15c-4903-b72d-e18c7077e888" Dec 01 08:55:29 crc kubenswrapper[4873]: I1201 08:55:29.051304 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-pfphw" event={"ID":"035095bd-1b65-4895-95a1-59feee524920","Type":"ContainerStarted","Data":"10cfa5ab3bf8be615873b6c88bd29df9bb7571e0acb9e53e9916d2c4cc45da9e"} Dec 01 08:55:29 crc kubenswrapper[4873]: I1201 08:55:29.052461 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-sjw48" event={"ID":"5ec1188a-c7f9-4cc8-8ea9-c1f1977041a5","Type":"ContainerStarted","Data":"5e6bcb931e6c4f7dcda6ecb422b1921aa1c34aa89f5b933d4ccc111fa2e66926"} Dec 01 08:55:29 crc kubenswrapper[4873]: I1201 08:55:29.053949 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-x9kc4" event={"ID":"450a3417-0074-4223-b1e3-aa1b854320fe","Type":"ContainerStarted","Data":"81e71bab0153c2fcbede9fdf426f226ea309258acf9fbbfe00dd72a14dc5d08b"} Dec 01 08:55:29 crc kubenswrapper[4873]: I1201 08:55:29.055578 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-n79q9" event={"ID":"7e3a2af8-3381-46e2-8c23-41aab8fd1a5e","Type":"ContainerStarted","Data":"f37561c71cc4e294cde8f00a1cb458b98f53b6a61a36b9bd46f627d8ea8acb48"} Dec 01 08:55:29 crc kubenswrapper[4873]: I1201 08:55:29.064501 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-j2vnt" event={"ID":"b45909ce-2ea4-4cf2-9351-c8839e44d734","Type":"ContainerStarted","Data":"6df59c48d7eaf0cfb17ace95a86158beca99b0a097eee8a326e61699dd3c1c7c"} Dec 01 08:55:29 crc kubenswrapper[4873]: I1201 08:55:29.069894 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6dxcw" event={"ID":"f7a7bcca-9403-4e2d-bc28-53d4eb5ae252","Type":"ContainerStarted","Data":"dcdba9c110970952cd1e95f87ea7c8afcc7edf95cf78f802c908f4fae65dbbea"} Dec 01 08:55:29 crc kubenswrapper[4873]: I1201 08:55:29.075282 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-49bz9" event={"ID":"879e02ea-306e-4e7b-9012-cb1cb6bdee00","Type":"ContainerStarted","Data":"8f27eb1e8c6811c325965d315b27976163da548f75c1f885f7f145d58ed3f04f"} Dec 01 08:55:29 crc kubenswrapper[4873]: E1201 08:55:29.078604 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-49bz9" podUID="879e02ea-306e-4e7b-9012-cb1cb6bdee00" Dec 01 08:55:29 crc kubenswrapper[4873]: I1201 08:55:29.078723 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-v7bvz" event={"ID":"3e0ad069-d971-4fd0-93df-b6a8794afe00","Type":"ContainerStarted","Data":"dfe8a85ab32135f577e64b9f0f284be9d717584f9249065eeba7cda2be71f5fd"} Dec 01 08:55:29 crc kubenswrapper[4873]: E1201 08:55:29.082372 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-v7bvz" podUID="3e0ad069-d971-4fd0-93df-b6a8794afe00" Dec 01 08:55:29 crc kubenswrapper[4873]: I1201 08:55:29.092987 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-dx4jm" event={"ID":"1cd3aa85-6333-4cfd-aacf-e51169ef4b42","Type":"ContainerStarted","Data":"56aa99b36acf6b96fac94bf33fd0d51edcbe12d27a51cf9e930a19ddf249d51a"} Dec 01 08:55:29 crc kubenswrapper[4873]: I1201 08:55:29.112484 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-d7c6w" event={"ID":"967215e8-7a18-4507-96c1-4c79c7e1d51a","Type":"ContainerStarted","Data":"48690f0cfe5aa24994cff477fe172f3a86dd0113e32d024d6d3cbdbb4bab9ecc"} Dec 01 08:55:29 crc kubenswrapper[4873]: I1201 08:55:29.114911 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-45whz" event={"ID":"94e7345b-b1ac-46e2-be25-9d64d3d33523","Type":"ContainerStarted","Data":"6936e2497a4b695fb425a86a6e169d3abc73ead0b37d8134d993b602c92ca1d7"} Dec 01 08:55:29 crc kubenswrapper[4873]: I1201 08:55:29.124050 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7wmqg" event={"ID":"eb53d542-63e9-487d-9d06-237c4b2b9252","Type":"ContainerStarted","Data":"d5681fdbe63438d7233e34bbe146a94f3f2b642bcde164cbb789867088560785"} Dec 01 08:55:29 crc kubenswrapper[4873]: I1201 08:55:29.365801 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/223a64ee-92cb-4b29-91bf-ffa7ed7d64ce-metrics-certs\") pod \"openstack-operator-controller-manager-7f4d6cf897-fddkc\" (UID: \"223a64ee-92cb-4b29-91bf-ffa7ed7d64ce\") " pod="openstack-operators/openstack-operator-controller-manager-7f4d6cf897-fddkc" Dec 01 08:55:29 crc kubenswrapper[4873]: I1201 08:55:29.365938 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/223a64ee-92cb-4b29-91bf-ffa7ed7d64ce-webhook-certs\") pod \"openstack-operator-controller-manager-7f4d6cf897-fddkc\" (UID: \"223a64ee-92cb-4b29-91bf-ffa7ed7d64ce\") " pod="openstack-operators/openstack-operator-controller-manager-7f4d6cf897-fddkc" Dec 01 08:55:29 crc kubenswrapper[4873]: E1201 08:55:29.366111 4873 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 01 08:55:29 crc kubenswrapper[4873]: E1201 08:55:29.366175 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/223a64ee-92cb-4b29-91bf-ffa7ed7d64ce-webhook-certs podName:223a64ee-92cb-4b29-91bf-ffa7ed7d64ce nodeName:}" failed. No retries permitted until 2025-12-01 08:55:33.366157041 +0000 UTC m=+909.268265580 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/223a64ee-92cb-4b29-91bf-ffa7ed7d64ce-webhook-certs") pod "openstack-operator-controller-manager-7f4d6cf897-fddkc" (UID: "223a64ee-92cb-4b29-91bf-ffa7ed7d64ce") : secret "webhook-server-cert" not found Dec 01 08:55:29 crc kubenswrapper[4873]: E1201 08:55:29.366619 4873 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 01 08:55:29 crc kubenswrapper[4873]: E1201 08:55:29.366810 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/223a64ee-92cb-4b29-91bf-ffa7ed7d64ce-metrics-certs podName:223a64ee-92cb-4b29-91bf-ffa7ed7d64ce nodeName:}" failed. No retries permitted until 2025-12-01 08:55:33.366788877 +0000 UTC m=+909.268897416 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/223a64ee-92cb-4b29-91bf-ffa7ed7d64ce-metrics-certs") pod "openstack-operator-controller-manager-7f4d6cf897-fddkc" (UID: "223a64ee-92cb-4b29-91bf-ffa7ed7d64ce") : secret "metrics-server-cert" not found Dec 01 08:55:30 crc kubenswrapper[4873]: E1201 08:55:30.272247 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cb9qn" podUID="664e20ef-b15c-4903-b72d-e18c7077e888" Dec 01 08:55:30 crc kubenswrapper[4873]: E1201 08:55:30.280278 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-49bz9" podUID="879e02ea-306e-4e7b-9012-cb1cb6bdee00" Dec 01 08:55:30 crc kubenswrapper[4873]: E1201 08:55:30.280364 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-v7bvz" podUID="3e0ad069-d971-4fd0-93df-b6a8794afe00" Dec 01 08:55:31 crc kubenswrapper[4873]: I1201 08:55:31.058704 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:55:31 crc kubenswrapper[4873]: I1201 08:55:31.058798 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:55:32 crc kubenswrapper[4873]: I1201 08:55:32.449186 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4dcc939e-b548-4fb6-814a-30e2aaa8a94a-cert\") pod \"infra-operator-controller-manager-57548d458d-tmdj7\" (UID: \"4dcc939e-b548-4fb6-814a-30e2aaa8a94a\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-tmdj7" Dec 01 08:55:32 crc kubenswrapper[4873]: E1201 08:55:32.449835 4873 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 01 08:55:32 crc kubenswrapper[4873]: E1201 08:55:32.449900 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4dcc939e-b548-4fb6-814a-30e2aaa8a94a-cert podName:4dcc939e-b548-4fb6-814a-30e2aaa8a94a nodeName:}" failed. No retries permitted until 2025-12-01 08:55:40.449882653 +0000 UTC m=+916.351991192 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/4dcc939e-b548-4fb6-814a-30e2aaa8a94a-cert") pod "infra-operator-controller-manager-57548d458d-tmdj7" (UID: "4dcc939e-b548-4fb6-814a-30e2aaa8a94a") : secret "infra-operator-webhook-server-cert" not found Dec 01 08:55:32 crc kubenswrapper[4873]: I1201 08:55:32.765193 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4vf8qt\" (UID: \"5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vf8qt" Dec 01 08:55:32 crc kubenswrapper[4873]: E1201 08:55:32.765659 4873 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 08:55:32 crc kubenswrapper[4873]: E1201 08:55:32.765748 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5-cert podName:5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5 nodeName:}" failed. No retries permitted until 2025-12-01 08:55:40.76572626 +0000 UTC m=+916.667834799 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4vf8qt" (UID: "5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 08:55:33 crc kubenswrapper[4873]: I1201 08:55:33.397698 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/223a64ee-92cb-4b29-91bf-ffa7ed7d64ce-webhook-certs\") pod \"openstack-operator-controller-manager-7f4d6cf897-fddkc\" (UID: \"223a64ee-92cb-4b29-91bf-ffa7ed7d64ce\") " pod="openstack-operators/openstack-operator-controller-manager-7f4d6cf897-fddkc" Dec 01 08:55:33 crc kubenswrapper[4873]: I1201 08:55:33.397791 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/223a64ee-92cb-4b29-91bf-ffa7ed7d64ce-metrics-certs\") pod \"openstack-operator-controller-manager-7f4d6cf897-fddkc\" (UID: \"223a64ee-92cb-4b29-91bf-ffa7ed7d64ce\") " pod="openstack-operators/openstack-operator-controller-manager-7f4d6cf897-fddkc" Dec 01 08:55:33 crc kubenswrapper[4873]: E1201 08:55:33.398322 4873 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 01 08:55:33 crc kubenswrapper[4873]: E1201 08:55:33.398384 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/223a64ee-92cb-4b29-91bf-ffa7ed7d64ce-metrics-certs podName:223a64ee-92cb-4b29-91bf-ffa7ed7d64ce nodeName:}" failed. No retries permitted until 2025-12-01 08:55:41.398365738 +0000 UTC m=+917.300474277 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/223a64ee-92cb-4b29-91bf-ffa7ed7d64ce-metrics-certs") pod "openstack-operator-controller-manager-7f4d6cf897-fddkc" (UID: "223a64ee-92cb-4b29-91bf-ffa7ed7d64ce") : secret "metrics-server-cert" not found Dec 01 08:55:33 crc kubenswrapper[4873]: E1201 08:55:33.399781 4873 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 01 08:55:33 crc kubenswrapper[4873]: E1201 08:55:33.399900 4873 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/223a64ee-92cb-4b29-91bf-ffa7ed7d64ce-webhook-certs podName:223a64ee-92cb-4b29-91bf-ffa7ed7d64ce nodeName:}" failed. No retries permitted until 2025-12-01 08:55:41.399860035 +0000 UTC m=+917.301968574 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/223a64ee-92cb-4b29-91bf-ffa7ed7d64ce-webhook-certs") pod "openstack-operator-controller-manager-7f4d6cf897-fddkc" (UID: "223a64ee-92cb-4b29-91bf-ffa7ed7d64ce") : secret "webhook-server-cert" not found Dec 01 08:55:40 crc kubenswrapper[4873]: I1201 08:55:40.505419 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4dcc939e-b548-4fb6-814a-30e2aaa8a94a-cert\") pod \"infra-operator-controller-manager-57548d458d-tmdj7\" (UID: \"4dcc939e-b548-4fb6-814a-30e2aaa8a94a\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-tmdj7" Dec 01 08:55:40 crc kubenswrapper[4873]: I1201 08:55:40.525268 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4dcc939e-b548-4fb6-814a-30e2aaa8a94a-cert\") pod \"infra-operator-controller-manager-57548d458d-tmdj7\" (UID: \"4dcc939e-b548-4fb6-814a-30e2aaa8a94a\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-tmdj7" Dec 01 08:55:40 crc kubenswrapper[4873]: I1201 08:55:40.615107 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-ptl6k" Dec 01 08:55:40 crc kubenswrapper[4873]: I1201 08:55:40.622829 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-tmdj7" Dec 01 08:55:40 crc kubenswrapper[4873]: I1201 08:55:40.814029 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4vf8qt\" (UID: \"5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vf8qt" Dec 01 08:55:40 crc kubenswrapper[4873]: I1201 08:55:40.817416 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4vf8qt\" (UID: \"5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vf8qt" Dec 01 08:55:41 crc kubenswrapper[4873]: I1201 08:55:41.048774 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vf8qt" Dec 01 08:55:41 crc kubenswrapper[4873]: I1201 08:55:41.429377 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/223a64ee-92cb-4b29-91bf-ffa7ed7d64ce-metrics-certs\") pod \"openstack-operator-controller-manager-7f4d6cf897-fddkc\" (UID: \"223a64ee-92cb-4b29-91bf-ffa7ed7d64ce\") " pod="openstack-operators/openstack-operator-controller-manager-7f4d6cf897-fddkc" Dec 01 08:55:41 crc kubenswrapper[4873]: I1201 08:55:41.429542 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/223a64ee-92cb-4b29-91bf-ffa7ed7d64ce-webhook-certs\") pod \"openstack-operator-controller-manager-7f4d6cf897-fddkc\" (UID: \"223a64ee-92cb-4b29-91bf-ffa7ed7d64ce\") " pod="openstack-operators/openstack-operator-controller-manager-7f4d6cf897-fddkc" Dec 01 08:55:41 crc kubenswrapper[4873]: I1201 08:55:41.439196 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/223a64ee-92cb-4b29-91bf-ffa7ed7d64ce-webhook-certs\") pod \"openstack-operator-controller-manager-7f4d6cf897-fddkc\" (UID: \"223a64ee-92cb-4b29-91bf-ffa7ed7d64ce\") " pod="openstack-operators/openstack-operator-controller-manager-7f4d6cf897-fddkc" Dec 01 08:55:41 crc kubenswrapper[4873]: I1201 08:55:41.450552 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/223a64ee-92cb-4b29-91bf-ffa7ed7d64ce-metrics-certs\") pod \"openstack-operator-controller-manager-7f4d6cf897-fddkc\" (UID: \"223a64ee-92cb-4b29-91bf-ffa7ed7d64ce\") " pod="openstack-operators/openstack-operator-controller-manager-7f4d6cf897-fddkc" Dec 01 08:55:41 crc kubenswrapper[4873]: I1201 08:55:41.562625 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7f4d6cf897-fddkc" Dec 01 08:55:42 crc kubenswrapper[4873]: I1201 08:55:42.678385 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-z2r2m"] Dec 01 08:55:42 crc kubenswrapper[4873]: I1201 08:55:42.681332 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z2r2m" Dec 01 08:55:42 crc kubenswrapper[4873]: I1201 08:55:42.688166 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z2r2m"] Dec 01 08:55:42 crc kubenswrapper[4873]: I1201 08:55:42.868305 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8f3966d-9400-44c1-a3e0-f4236541480e-catalog-content\") pod \"redhat-operators-z2r2m\" (UID: \"f8f3966d-9400-44c1-a3e0-f4236541480e\") " pod="openshift-marketplace/redhat-operators-z2r2m" Dec 01 08:55:42 crc kubenswrapper[4873]: I1201 08:55:42.868447 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4c6jw\" (UniqueName: \"kubernetes.io/projected/f8f3966d-9400-44c1-a3e0-f4236541480e-kube-api-access-4c6jw\") pod \"redhat-operators-z2r2m\" (UID: \"f8f3966d-9400-44c1-a3e0-f4236541480e\") " pod="openshift-marketplace/redhat-operators-z2r2m" Dec 01 08:55:42 crc kubenswrapper[4873]: I1201 08:55:42.868512 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8f3966d-9400-44c1-a3e0-f4236541480e-utilities\") pod \"redhat-operators-z2r2m\" (UID: \"f8f3966d-9400-44c1-a3e0-f4236541480e\") " pod="openshift-marketplace/redhat-operators-z2r2m" Dec 01 08:55:42 crc kubenswrapper[4873]: I1201 08:55:42.970409 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4c6jw\" (UniqueName: \"kubernetes.io/projected/f8f3966d-9400-44c1-a3e0-f4236541480e-kube-api-access-4c6jw\") pod \"redhat-operators-z2r2m\" (UID: \"f8f3966d-9400-44c1-a3e0-f4236541480e\") " pod="openshift-marketplace/redhat-operators-z2r2m" Dec 01 08:55:42 crc kubenswrapper[4873]: I1201 08:55:42.970539 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8f3966d-9400-44c1-a3e0-f4236541480e-utilities\") pod \"redhat-operators-z2r2m\" (UID: \"f8f3966d-9400-44c1-a3e0-f4236541480e\") " pod="openshift-marketplace/redhat-operators-z2r2m" Dec 01 08:55:42 crc kubenswrapper[4873]: I1201 08:55:42.970597 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8f3966d-9400-44c1-a3e0-f4236541480e-catalog-content\") pod \"redhat-operators-z2r2m\" (UID: \"f8f3966d-9400-44c1-a3e0-f4236541480e\") " pod="openshift-marketplace/redhat-operators-z2r2m" Dec 01 08:55:42 crc kubenswrapper[4873]: I1201 08:55:42.971194 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8f3966d-9400-44c1-a3e0-f4236541480e-catalog-content\") pod \"redhat-operators-z2r2m\" (UID: \"f8f3966d-9400-44c1-a3e0-f4236541480e\") " pod="openshift-marketplace/redhat-operators-z2r2m" Dec 01 08:55:42 crc kubenswrapper[4873]: I1201 08:55:42.971303 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8f3966d-9400-44c1-a3e0-f4236541480e-utilities\") pod \"redhat-operators-z2r2m\" (UID: \"f8f3966d-9400-44c1-a3e0-f4236541480e\") " pod="openshift-marketplace/redhat-operators-z2r2m" Dec 01 08:55:42 crc kubenswrapper[4873]: I1201 08:55:42.992741 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4c6jw\" (UniqueName: \"kubernetes.io/projected/f8f3966d-9400-44c1-a3e0-f4236541480e-kube-api-access-4c6jw\") pod \"redhat-operators-z2r2m\" (UID: \"f8f3966d-9400-44c1-a3e0-f4236541480e\") " pod="openshift-marketplace/redhat-operators-z2r2m" Dec 01 08:55:43 crc kubenswrapper[4873]: I1201 08:55:43.032594 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z2r2m" Dec 01 08:55:45 crc kubenswrapper[4873]: E1201 08:55:45.814466 4873 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/barbican-operator@sha256:f6059a0fbf031d34dcf086d14ce8c0546caeaee23c5780e90b5037c5feee9fea" Dec 01 08:55:45 crc kubenswrapper[4873]: E1201 08:55:45.815432 4873 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/barbican-operator@sha256:f6059a0fbf031d34dcf086d14ce8c0546caeaee23c5780e90b5037c5feee9fea,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-t9hnx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-7d9dfd778-5rq8z_openstack-operators(51253ba6-e0b5-44ac-8c18-be17b4d13024): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 08:55:46 crc kubenswrapper[4873]: E1201 08:55:46.632024 4873 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5" Dec 01 08:55:46 crc kubenswrapper[4873]: E1201 08:55:46.632956 4873 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qgqgk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-68c6d99b8f-45whz_openstack-operators(94e7345b-b1ac-46e2-be25-9d64d3d33523): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 08:55:47 crc kubenswrapper[4873]: E1201 08:55:47.231713 4873 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168" Dec 01 08:55:47 crc kubenswrapper[4873]: E1201 08:55:47.232006 4873 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9hnhn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-n79q9_openstack-operators(7e3a2af8-3381-46e2-8c23-41aab8fd1a5e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 08:55:49 crc kubenswrapper[4873]: I1201 08:55:49.993391 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tdc2x"] Dec 01 08:55:49 crc kubenswrapper[4873]: I1201 08:55:49.995846 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tdc2x" Dec 01 08:55:50 crc kubenswrapper[4873]: I1201 08:55:50.085669 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77cdace7-aa56-4a6b-900b-b162ef0410a4-utilities\") pod \"certified-operators-tdc2x\" (UID: \"77cdace7-aa56-4a6b-900b-b162ef0410a4\") " pod="openshift-marketplace/certified-operators-tdc2x" Dec 01 08:55:50 crc kubenswrapper[4873]: I1201 08:55:50.085785 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnlk5\" (UniqueName: \"kubernetes.io/projected/77cdace7-aa56-4a6b-900b-b162ef0410a4-kube-api-access-rnlk5\") pod \"certified-operators-tdc2x\" (UID: \"77cdace7-aa56-4a6b-900b-b162ef0410a4\") " pod="openshift-marketplace/certified-operators-tdc2x" Dec 01 08:55:50 crc kubenswrapper[4873]: I1201 08:55:50.086096 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77cdace7-aa56-4a6b-900b-b162ef0410a4-catalog-content\") pod \"certified-operators-tdc2x\" (UID: \"77cdace7-aa56-4a6b-900b-b162ef0410a4\") " pod="openshift-marketplace/certified-operators-tdc2x" Dec 01 08:55:50 crc kubenswrapper[4873]: I1201 08:55:50.107463 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tdc2x"] Dec 01 08:55:50 crc kubenswrapper[4873]: I1201 08:55:50.188046 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77cdace7-aa56-4a6b-900b-b162ef0410a4-catalog-content\") pod \"certified-operators-tdc2x\" (UID: \"77cdace7-aa56-4a6b-900b-b162ef0410a4\") " pod="openshift-marketplace/certified-operators-tdc2x" Dec 01 08:55:50 crc kubenswrapper[4873]: I1201 08:55:50.188552 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77cdace7-aa56-4a6b-900b-b162ef0410a4-utilities\") pod \"certified-operators-tdc2x\" (UID: \"77cdace7-aa56-4a6b-900b-b162ef0410a4\") " pod="openshift-marketplace/certified-operators-tdc2x" Dec 01 08:55:50 crc kubenswrapper[4873]: I1201 08:55:50.188623 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnlk5\" (UniqueName: \"kubernetes.io/projected/77cdace7-aa56-4a6b-900b-b162ef0410a4-kube-api-access-rnlk5\") pod \"certified-operators-tdc2x\" (UID: \"77cdace7-aa56-4a6b-900b-b162ef0410a4\") " pod="openshift-marketplace/certified-operators-tdc2x" Dec 01 08:55:50 crc kubenswrapper[4873]: I1201 08:55:50.188735 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77cdace7-aa56-4a6b-900b-b162ef0410a4-catalog-content\") pod \"certified-operators-tdc2x\" (UID: \"77cdace7-aa56-4a6b-900b-b162ef0410a4\") " pod="openshift-marketplace/certified-operators-tdc2x" Dec 01 08:55:50 crc kubenswrapper[4873]: I1201 08:55:50.189196 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77cdace7-aa56-4a6b-900b-b162ef0410a4-utilities\") pod \"certified-operators-tdc2x\" (UID: \"77cdace7-aa56-4a6b-900b-b162ef0410a4\") " pod="openshift-marketplace/certified-operators-tdc2x" Dec 01 08:55:50 crc kubenswrapper[4873]: I1201 08:55:50.230375 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnlk5\" (UniqueName: \"kubernetes.io/projected/77cdace7-aa56-4a6b-900b-b162ef0410a4-kube-api-access-rnlk5\") pod \"certified-operators-tdc2x\" (UID: \"77cdace7-aa56-4a6b-900b-b162ef0410a4\") " pod="openshift-marketplace/certified-operators-tdc2x" Dec 01 08:55:50 crc kubenswrapper[4873]: I1201 08:55:50.434212 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tdc2x" Dec 01 08:55:51 crc kubenswrapper[4873]: E1201 08:55:51.041601 4873 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:0f523b7e2fa9e86fef986acf07d0c42d5658c475d565f11eaea926ebffcb6530" Dec 01 08:55:51 crc kubenswrapper[4873]: E1201 08:55:51.041865 4873 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:0f523b7e2fa9e86fef986acf07d0c42d5658c475d565f11eaea926ebffcb6530,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6phsr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-6c548fd776-7wmqg_openstack-operators(eb53d542-63e9-487d-9d06-237c4b2b9252): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 08:55:51 crc kubenswrapper[4873]: E1201 08:55:51.167291 4873 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.103:5001/openstack-k8s-operators/glance-operator:c8bcc1d6cd56194574b4d200ed5452dd2f1e0ab0" Dec 01 08:55:51 crc kubenswrapper[4873]: E1201 08:55:51.167368 4873 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.103:5001/openstack-k8s-operators/glance-operator:c8bcc1d6cd56194574b4d200ed5452dd2f1e0ab0" Dec 01 08:55:51 crc kubenswrapper[4873]: E1201 08:55:51.167548 4873 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.103:5001/openstack-k8s-operators/glance-operator:c8bcc1d6cd56194574b4d200ed5452dd2f1e0ab0,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jbjxp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-5cd474b4ff-wcmv4_openstack-operators(f778fd6c-4c8a-4067-b3f2-cb7d98a50bf7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 08:55:52 crc kubenswrapper[4873]: E1201 08:55:52.453611 4873 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385" Dec 01 08:55:52 crc kubenswrapper[4873]: E1201 08:55:52.454172 4873 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rmb2k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-bhrpn_openstack-operators(225323c8-ed2f-4573-920d-f43f3e4561de): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 08:55:53 crc kubenswrapper[4873]: E1201 08:55:53.308646 4873 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:ecf7be921850bdc04697ed1b332bab39ad2a64e4e45c2a445c04f9bae6ac61b5" Dec 01 08:55:53 crc kubenswrapper[4873]: E1201 08:55:53.308948 4873 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:ecf7be921850bdc04697ed1b332bab39ad2a64e4e45c2a445c04f9bae6ac61b5,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-k7tkm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-6546668bfd-j2vnt_openstack-operators(b45909ce-2ea4-4cf2-9351-c8839e44d734): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 08:55:53 crc kubenswrapper[4873]: I1201 08:55:53.333981 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-f6n4k"] Dec 01 08:55:53 crc kubenswrapper[4873]: I1201 08:55:53.336949 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f6n4k" Dec 01 08:55:53 crc kubenswrapper[4873]: I1201 08:55:53.351057 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f6n4k"] Dec 01 08:55:53 crc kubenswrapper[4873]: I1201 08:55:53.457420 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/074c5f85-694d-47af-a33f-e1995ae73a5c-utilities\") pod \"redhat-marketplace-f6n4k\" (UID: \"074c5f85-694d-47af-a33f-e1995ae73a5c\") " pod="openshift-marketplace/redhat-marketplace-f6n4k" Dec 01 08:55:53 crc kubenswrapper[4873]: I1201 08:55:53.457466 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkgsl\" (UniqueName: \"kubernetes.io/projected/074c5f85-694d-47af-a33f-e1995ae73a5c-kube-api-access-gkgsl\") pod \"redhat-marketplace-f6n4k\" (UID: \"074c5f85-694d-47af-a33f-e1995ae73a5c\") " pod="openshift-marketplace/redhat-marketplace-f6n4k" Dec 01 08:55:53 crc kubenswrapper[4873]: I1201 08:55:53.457564 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/074c5f85-694d-47af-a33f-e1995ae73a5c-catalog-content\") pod \"redhat-marketplace-f6n4k\" (UID: \"074c5f85-694d-47af-a33f-e1995ae73a5c\") " pod="openshift-marketplace/redhat-marketplace-f6n4k" Dec 01 08:55:53 crc kubenswrapper[4873]: I1201 08:55:53.559093 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/074c5f85-694d-47af-a33f-e1995ae73a5c-catalog-content\") pod \"redhat-marketplace-f6n4k\" (UID: \"074c5f85-694d-47af-a33f-e1995ae73a5c\") " pod="openshift-marketplace/redhat-marketplace-f6n4k" Dec 01 08:55:53 crc kubenswrapper[4873]: I1201 08:55:53.559168 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/074c5f85-694d-47af-a33f-e1995ae73a5c-utilities\") pod \"redhat-marketplace-f6n4k\" (UID: \"074c5f85-694d-47af-a33f-e1995ae73a5c\") " pod="openshift-marketplace/redhat-marketplace-f6n4k" Dec 01 08:55:53 crc kubenswrapper[4873]: I1201 08:55:53.559196 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkgsl\" (UniqueName: \"kubernetes.io/projected/074c5f85-694d-47af-a33f-e1995ae73a5c-kube-api-access-gkgsl\") pod \"redhat-marketplace-f6n4k\" (UID: \"074c5f85-694d-47af-a33f-e1995ae73a5c\") " pod="openshift-marketplace/redhat-marketplace-f6n4k" Dec 01 08:55:53 crc kubenswrapper[4873]: I1201 08:55:53.560031 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/074c5f85-694d-47af-a33f-e1995ae73a5c-utilities\") pod \"redhat-marketplace-f6n4k\" (UID: \"074c5f85-694d-47af-a33f-e1995ae73a5c\") " pod="openshift-marketplace/redhat-marketplace-f6n4k" Dec 01 08:55:53 crc kubenswrapper[4873]: I1201 08:55:53.560170 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/074c5f85-694d-47af-a33f-e1995ae73a5c-catalog-content\") pod \"redhat-marketplace-f6n4k\" (UID: \"074c5f85-694d-47af-a33f-e1995ae73a5c\") " pod="openshift-marketplace/redhat-marketplace-f6n4k" Dec 01 08:55:53 crc kubenswrapper[4873]: I1201 08:55:53.582164 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkgsl\" (UniqueName: \"kubernetes.io/projected/074c5f85-694d-47af-a33f-e1995ae73a5c-kube-api-access-gkgsl\") pod \"redhat-marketplace-f6n4k\" (UID: \"074c5f85-694d-47af-a33f-e1995ae73a5c\") " pod="openshift-marketplace/redhat-marketplace-f6n4k" Dec 01 08:55:53 crc kubenswrapper[4873]: I1201 08:55:53.699744 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f6n4k" Dec 01 08:55:54 crc kubenswrapper[4873]: E1201 08:55:54.227471 4873 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621" Dec 01 08:55:54 crc kubenswrapper[4873]: E1201 08:55:54.227749 4873 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-994kd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-dx4jm_openstack-operators(1cd3aa85-6333-4cfd-aacf-e51169ef4b42): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 08:55:54 crc kubenswrapper[4873]: E1201 08:55:54.872884 4873 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7" Dec 01 08:55:54 crc kubenswrapper[4873]: E1201 08:55:54.873502 4873 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6vlf2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-56bbcc9d85-hnzzf_openstack-operators(3eff7ce2-9ff7-413c-b472-9e114e7130ca): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 08:55:55 crc kubenswrapper[4873]: E1201 08:55:55.422266 4873 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670" Dec 01 08:55:55 crc kubenswrapper[4873]: E1201 08:55:55.422482 4873 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-j9k7x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-x9kc4_openstack-operators(450a3417-0074-4223-b1e3-aa1b854320fe): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 08:55:55 crc kubenswrapper[4873]: I1201 08:55:55.661086 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jvrlm"] Dec 01 08:55:55 crc kubenswrapper[4873]: I1201 08:55:55.687710 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jvrlm" Dec 01 08:55:55 crc kubenswrapper[4873]: I1201 08:55:55.698379 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jvrlm"] Dec 01 08:55:55 crc kubenswrapper[4873]: I1201 08:55:55.811006 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdrp8\" (UniqueName: \"kubernetes.io/projected/439ab38e-6a5b-4aff-baef-0be711a2fad5-kube-api-access-kdrp8\") pod \"community-operators-jvrlm\" (UID: \"439ab38e-6a5b-4aff-baef-0be711a2fad5\") " pod="openshift-marketplace/community-operators-jvrlm" Dec 01 08:55:55 crc kubenswrapper[4873]: I1201 08:55:55.811902 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/439ab38e-6a5b-4aff-baef-0be711a2fad5-utilities\") pod \"community-operators-jvrlm\" (UID: \"439ab38e-6a5b-4aff-baef-0be711a2fad5\") " pod="openshift-marketplace/community-operators-jvrlm" Dec 01 08:55:55 crc kubenswrapper[4873]: I1201 08:55:55.811948 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/439ab38e-6a5b-4aff-baef-0be711a2fad5-catalog-content\") pod \"community-operators-jvrlm\" (UID: \"439ab38e-6a5b-4aff-baef-0be711a2fad5\") " pod="openshift-marketplace/community-operators-jvrlm" Dec 01 08:55:55 crc kubenswrapper[4873]: I1201 08:55:55.914706 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/439ab38e-6a5b-4aff-baef-0be711a2fad5-utilities\") pod \"community-operators-jvrlm\" (UID: \"439ab38e-6a5b-4aff-baef-0be711a2fad5\") " pod="openshift-marketplace/community-operators-jvrlm" Dec 01 08:55:55 crc kubenswrapper[4873]: I1201 08:55:55.914768 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/439ab38e-6a5b-4aff-baef-0be711a2fad5-catalog-content\") pod \"community-operators-jvrlm\" (UID: \"439ab38e-6a5b-4aff-baef-0be711a2fad5\") " pod="openshift-marketplace/community-operators-jvrlm" Dec 01 08:55:55 crc kubenswrapper[4873]: I1201 08:55:55.914907 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdrp8\" (UniqueName: \"kubernetes.io/projected/439ab38e-6a5b-4aff-baef-0be711a2fad5-kube-api-access-kdrp8\") pod \"community-operators-jvrlm\" (UID: \"439ab38e-6a5b-4aff-baef-0be711a2fad5\") " pod="openshift-marketplace/community-operators-jvrlm" Dec 01 08:55:55 crc kubenswrapper[4873]: I1201 08:55:55.915818 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/439ab38e-6a5b-4aff-baef-0be711a2fad5-utilities\") pod \"community-operators-jvrlm\" (UID: \"439ab38e-6a5b-4aff-baef-0be711a2fad5\") " pod="openshift-marketplace/community-operators-jvrlm" Dec 01 08:55:55 crc kubenswrapper[4873]: I1201 08:55:55.916119 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/439ab38e-6a5b-4aff-baef-0be711a2fad5-catalog-content\") pod \"community-operators-jvrlm\" (UID: \"439ab38e-6a5b-4aff-baef-0be711a2fad5\") " pod="openshift-marketplace/community-operators-jvrlm" Dec 01 08:55:55 crc kubenswrapper[4873]: I1201 08:55:55.940799 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdrp8\" (UniqueName: \"kubernetes.io/projected/439ab38e-6a5b-4aff-baef-0be711a2fad5-kube-api-access-kdrp8\") pod \"community-operators-jvrlm\" (UID: \"439ab38e-6a5b-4aff-baef-0be711a2fad5\") " pod="openshift-marketplace/community-operators-jvrlm" Dec 01 08:55:56 crc kubenswrapper[4873]: I1201 08:55:56.017024 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jvrlm" Dec 01 08:55:56 crc kubenswrapper[4873]: E1201 08:55:56.464936 4873 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:986861e5a0a9954f63581d9d55a30f8057883cefea489415d76257774526eea3" Dec 01 08:55:56 crc kubenswrapper[4873]: E1201 08:55:56.465262 4873 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:986861e5a0a9954f63581d9d55a30f8057883cefea489415d76257774526eea3,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8xtkm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-546d4bdf48-pfphw_openstack-operators(035095bd-1b65-4895-95a1-59feee524920): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 08:55:59 crc kubenswrapper[4873]: I1201 08:55:59.862022 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-tmdj7"] Dec 01 08:56:01 crc kubenswrapper[4873]: I1201 08:56:01.035890 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z2r2m"] Dec 01 08:56:01 crc kubenswrapper[4873]: I1201 08:56:01.059137 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:56:01 crc kubenswrapper[4873]: I1201 08:56:01.059201 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:56:01 crc kubenswrapper[4873]: I1201 08:56:01.059300 4873 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" Dec 01 08:56:01 crc kubenswrapper[4873]: I1201 08:56:01.060345 4873 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2af36981211968ab2b3e6fc266a51d247e83e9c0c856f6566cda62f7eeb85b9a"} pod="openshift-machine-config-operator/machine-config-daemon-scwpp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 08:56:01 crc kubenswrapper[4873]: I1201 08:56:01.060444 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" containerID="cri-o://2af36981211968ab2b3e6fc266a51d247e83e9c0c856f6566cda62f7eeb85b9a" gracePeriod=600 Dec 01 08:56:01 crc kubenswrapper[4873]: I1201 08:56:01.798184 4873 generic.go:334] "Generic (PLEG): container finished" podID="fef7b114-0e07-402d-a37b-315c36011f4b" containerID="2af36981211968ab2b3e6fc266a51d247e83e9c0c856f6566cda62f7eeb85b9a" exitCode=0 Dec 01 08:56:01 crc kubenswrapper[4873]: I1201 08:56:01.798262 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" event={"ID":"fef7b114-0e07-402d-a37b-315c36011f4b","Type":"ContainerDied","Data":"2af36981211968ab2b3e6fc266a51d247e83e9c0c856f6566cda62f7eeb85b9a"} Dec 01 08:56:01 crc kubenswrapper[4873]: I1201 08:56:01.798881 4873 scope.go:117] "RemoveContainer" containerID="6650a48ef92f241e9c56ee0be1ecde624a461ff74d058cf63149c0cd17ae8d2d" Dec 01 08:56:01 crc kubenswrapper[4873]: E1201 08:56:01.919711 4873 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 01 08:56:01 crc kubenswrapper[4873]: E1201 08:56:01.920110 4873 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nrqtf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-6dxcw_openstack-operators(f7a7bcca-9403-4e2d-bc28-53d4eb5ae252): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 08:56:01 crc kubenswrapper[4873]: E1201 08:56:01.921456 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6dxcw" podUID="f7a7bcca-9403-4e2d-bc28-53d4eb5ae252" Dec 01 08:56:01 crc kubenswrapper[4873]: W1201 08:56:01.982941 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4dcc939e_b548_4fb6_814a_30e2aaa8a94a.slice/crio-365aa0a1ebf94d39b81ce91eb33a8c2b1690f671d860ec0bee38848c48c1fc55 WatchSource:0}: Error finding container 365aa0a1ebf94d39b81ce91eb33a8c2b1690f671d860ec0bee38848c48c1fc55: Status 404 returned error can't find the container with id 365aa0a1ebf94d39b81ce91eb33a8c2b1690f671d860ec0bee38848c48c1fc55 Dec 01 08:56:02 crc kubenswrapper[4873]: W1201 08:56:02.010658 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf8f3966d_9400_44c1_a3e0_f4236541480e.slice/crio-dfca0c024b618172217edc647a9e04be8051e8782de00bba4e8281987d978456 WatchSource:0}: Error finding container dfca0c024b618172217edc647a9e04be8051e8782de00bba4e8281987d978456: Status 404 returned error can't find the container with id dfca0c024b618172217edc647a9e04be8051e8782de00bba4e8281987d978456 Dec 01 08:56:02 crc kubenswrapper[4873]: I1201 08:56:02.652815 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7f4d6cf897-fddkc"] Dec 01 08:56:02 crc kubenswrapper[4873]: I1201 08:56:02.812627 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-mvc26" event={"ID":"34e3726b-c9f5-4a86-9e87-e179f8dae739","Type":"ContainerStarted","Data":"3fa7b2688c6bc5dc7f598c73b8d15eea30c1a33c44372ee95dfe9141cb93936a"} Dec 01 08:56:02 crc kubenswrapper[4873]: I1201 08:56:02.821339 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-gqm5w" event={"ID":"671ad3cc-9716-42de-b5ae-fb69847d3bd7","Type":"ContainerStarted","Data":"e6b8d7672fac8e4e8020470e0570ab6d1b6ccaadbd136589d8d51f263b6f6df2"} Dec 01 08:56:02 crc kubenswrapper[4873]: I1201 08:56:02.823536 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-sjw48" event={"ID":"5ec1188a-c7f9-4cc8-8ea9-c1f1977041a5","Type":"ContainerStarted","Data":"02adc01b7b71625e4632eba8197f9d3daa5b6d9e2c94f954ad9e59ccdb3f83dd"} Dec 01 08:56:02 crc kubenswrapper[4873]: I1201 08:56:02.825497 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-d7c6w" event={"ID":"967215e8-7a18-4507-96c1-4c79c7e1d51a","Type":"ContainerStarted","Data":"6fd9a13a99d0a4527226780872259f3dbfc3a4c2c7fa4adb5e312d3a944606fe"} Dec 01 08:56:02 crc kubenswrapper[4873]: I1201 08:56:02.829500 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-6r8hb" event={"ID":"723f446d-21bb-432a-b2af-93219593819d","Type":"ContainerStarted","Data":"390e230445cdb0dc8245815dd32cc6dfb1e44549242cccbdf9ca13236a244fe3"} Dec 01 08:56:02 crc kubenswrapper[4873]: I1201 08:56:02.830446 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z2r2m" event={"ID":"f8f3966d-9400-44c1-a3e0-f4236541480e","Type":"ContainerStarted","Data":"dfca0c024b618172217edc647a9e04be8051e8782de00bba4e8281987d978456"} Dec 01 08:56:02 crc kubenswrapper[4873]: I1201 08:56:02.831386 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7f4d6cf897-fddkc" event={"ID":"223a64ee-92cb-4b29-91bf-ffa7ed7d64ce","Type":"ContainerStarted","Data":"086f2102538f6d934ba876c359bdf865673ce0c1423bf912c056eebd0fa0e62f"} Dec 01 08:56:02 crc kubenswrapper[4873]: I1201 08:56:02.833667 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-tmdj7" event={"ID":"4dcc939e-b548-4fb6-814a-30e2aaa8a94a","Type":"ContainerStarted","Data":"365aa0a1ebf94d39b81ce91eb33a8c2b1690f671d860ec0bee38848c48c1fc55"} Dec 01 08:56:03 crc kubenswrapper[4873]: I1201 08:56:03.190520 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vf8qt"] Dec 01 08:56:03 crc kubenswrapper[4873]: I1201 08:56:03.195049 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jvrlm"] Dec 01 08:56:03 crc kubenswrapper[4873]: I1201 08:56:03.295550 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tdc2x"] Dec 01 08:56:03 crc kubenswrapper[4873]: W1201 08:56:03.311144 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod439ab38e_6a5b_4aff_baef_0be711a2fad5.slice/crio-c2d37e795ab7413bbc0160cf511a0bf8ce9e930f4abfa1bcd5ac8654cbccb517 WatchSource:0}: Error finding container c2d37e795ab7413bbc0160cf511a0bf8ce9e930f4abfa1bcd5ac8654cbccb517: Status 404 returned error can't find the container with id c2d37e795ab7413bbc0160cf511a0bf8ce9e930f4abfa1bcd5ac8654cbccb517 Dec 01 08:56:03 crc kubenswrapper[4873]: I1201 08:56:03.314222 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f6n4k"] Dec 01 08:56:03 crc kubenswrapper[4873]: W1201 08:56:03.361034 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod77cdace7_aa56_4a6b_900b_b162ef0410a4.slice/crio-0aa3a2f837ad59345f1a5a1b26d886cfb7d43963b971f621a2e0c5295777b1da WatchSource:0}: Error finding container 0aa3a2f837ad59345f1a5a1b26d886cfb7d43963b971f621a2e0c5295777b1da: Status 404 returned error can't find the container with id 0aa3a2f837ad59345f1a5a1b26d886cfb7d43963b971f621a2e0c5295777b1da Dec 01 08:56:03 crc kubenswrapper[4873]: E1201 08:56:03.363714 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-j2vnt" podUID="b45909ce-2ea4-4cf2-9351-c8839e44d734" Dec 01 08:56:03 crc kubenswrapper[4873]: W1201 08:56:03.370252 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod074c5f85_694d_47af_a33f_e1995ae73a5c.slice/crio-1fc588e35318ef8b19cd182fddcefe703211662ebe652497c89196c3191cfe7e WatchSource:0}: Error finding container 1fc588e35318ef8b19cd182fddcefe703211662ebe652497c89196c3191cfe7e: Status 404 returned error can't find the container with id 1fc588e35318ef8b19cd182fddcefe703211662ebe652497c89196c3191cfe7e Dec 01 08:56:03 crc kubenswrapper[4873]: E1201 08:56:03.401683 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-x9kc4" podUID="450a3417-0074-4223-b1e3-aa1b854320fe" Dec 01 08:56:03 crc kubenswrapper[4873]: I1201 08:56:03.847245 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vf8qt" event={"ID":"5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5","Type":"ContainerStarted","Data":"ebfa2d72d7b0dfd8175cfd7ad556413fc50baa3f57fcdb9fa70e442c21618af3"} Dec 01 08:56:03 crc kubenswrapper[4873]: I1201 08:56:03.858497 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f6n4k" event={"ID":"074c5f85-694d-47af-a33f-e1995ae73a5c","Type":"ContainerStarted","Data":"1fc588e35318ef8b19cd182fddcefe703211662ebe652497c89196c3191cfe7e"} Dec 01 08:56:03 crc kubenswrapper[4873]: I1201 08:56:03.869671 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-v7bvz" event={"ID":"3e0ad069-d971-4fd0-93df-b6a8794afe00","Type":"ContainerStarted","Data":"dc84e6f9f6c18bc4a31cc4fe4946d7693b58c1f0cd276ef8107cdd18097b32d1"} Dec 01 08:56:03 crc kubenswrapper[4873]: I1201 08:56:03.873496 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jvrlm" event={"ID":"439ab38e-6a5b-4aff-baef-0be711a2fad5","Type":"ContainerStarted","Data":"c2d37e795ab7413bbc0160cf511a0bf8ce9e930f4abfa1bcd5ac8654cbccb517"} Dec 01 08:56:03 crc kubenswrapper[4873]: I1201 08:56:03.876613 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-j2vnt" event={"ID":"b45909ce-2ea4-4cf2-9351-c8839e44d734","Type":"ContainerStarted","Data":"73862d0bdd4e7dd4aca9e10df9ec530fe5d28978296611114a5ae911e2a4cead"} Dec 01 08:56:03 crc kubenswrapper[4873]: I1201 08:56:03.882647 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cb9qn" event={"ID":"664e20ef-b15c-4903-b72d-e18c7077e888","Type":"ContainerStarted","Data":"77b67adcde9e8fab58588a7bd31d0de47591039787f3449586476bbac3918056"} Dec 01 08:56:03 crc kubenswrapper[4873]: I1201 08:56:03.889209 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-49bz9" event={"ID":"879e02ea-306e-4e7b-9012-cb1cb6bdee00","Type":"ContainerStarted","Data":"d83ba024c525df6640ffd57acd903d6ae7d908312b0b7c1f5d5b856b1b2ed69f"} Dec 01 08:56:03 crc kubenswrapper[4873]: I1201 08:56:03.911441 4873 generic.go:334] "Generic (PLEG): container finished" podID="f8f3966d-9400-44c1-a3e0-f4236541480e" containerID="6a839b7303a1d01b67d8aca54815c2aa26443339aa86144c0a21b02b3ccac472" exitCode=0 Dec 01 08:56:03 crc kubenswrapper[4873]: I1201 08:56:03.911545 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z2r2m" event={"ID":"f8f3966d-9400-44c1-a3e0-f4236541480e","Type":"ContainerDied","Data":"6a839b7303a1d01b67d8aca54815c2aa26443339aa86144c0a21b02b3ccac472"} Dec 01 08:56:03 crc kubenswrapper[4873]: I1201 08:56:03.918713 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tdc2x" event={"ID":"77cdace7-aa56-4a6b-900b-b162ef0410a4","Type":"ContainerStarted","Data":"0aa3a2f837ad59345f1a5a1b26d886cfb7d43963b971f621a2e0c5295777b1da"} Dec 01 08:56:03 crc kubenswrapper[4873]: I1201 08:56:03.927961 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cb9qn" podStartSLOduration=6.294100115 podStartE2EDuration="39.927941295s" podCreationTimestamp="2025-12-01 08:55:24 +0000 UTC" firstStartedPulling="2025-12-01 08:55:28.670562468 +0000 UTC m=+904.572670997" lastFinishedPulling="2025-12-01 08:56:02.304403638 +0000 UTC m=+938.206512177" observedRunningTime="2025-12-01 08:56:03.917982437 +0000 UTC m=+939.820090976" watchObservedRunningTime="2025-12-01 08:56:03.927941295 +0000 UTC m=+939.830049834" Dec 01 08:56:03 crc kubenswrapper[4873]: I1201 08:56:03.929837 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" event={"ID":"fef7b114-0e07-402d-a37b-315c36011f4b","Type":"ContainerStarted","Data":"0976ae6cae359bbc0353d195c023558e0d48d039655f0dbf281e76be957bb9b6"} Dec 01 08:56:03 crc kubenswrapper[4873]: I1201 08:56:03.955636 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-x9kc4" event={"ID":"450a3417-0074-4223-b1e3-aa1b854320fe","Type":"ContainerStarted","Data":"6cda21298a06b0746bc7bf190f3c4e6674db4803ef83c9aa50967df545cf235e"} Dec 01 08:56:03 crc kubenswrapper[4873]: E1201 08:56:03.960249 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670\\\"\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-x9kc4" podUID="450a3417-0074-4223-b1e3-aa1b854320fe" Dec 01 08:56:04 crc kubenswrapper[4873]: E1201 08:56:04.329294 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-45whz" podUID="94e7345b-b1ac-46e2-be25-9d64d3d33523" Dec 01 08:56:04 crc kubenswrapper[4873]: E1201 08:56:04.499592 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-5cd474b4ff-wcmv4" podUID="f778fd6c-4c8a-4067-b3f2-cb7d98a50bf7" Dec 01 08:56:04 crc kubenswrapper[4873]: E1201 08:56:04.581361 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-bhrpn" podUID="225323c8-ed2f-4573-920d-f43f3e4561de" Dec 01 08:56:04 crc kubenswrapper[4873]: E1201 08:56:04.928142 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-998648c74-n79q9" podUID="7e3a2af8-3381-46e2-8c23-41aab8fd1a5e" Dec 01 08:56:04 crc kubenswrapper[4873]: I1201 08:56:04.971693 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-gqm5w" event={"ID":"671ad3cc-9716-42de-b5ae-fb69847d3bd7","Type":"ContainerStarted","Data":"c00c44c5d242a49422d1f742080ab92bc99e5e09b563315d28ff112608782510"} Dec 01 08:56:04 crc kubenswrapper[4873]: I1201 08:56:04.972208 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-gqm5w" Dec 01 08:56:04 crc kubenswrapper[4873]: I1201 08:56:04.981601 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5cd474b4ff-wcmv4" event={"ID":"f778fd6c-4c8a-4067-b3f2-cb7d98a50bf7","Type":"ContainerStarted","Data":"b0661e85a63d569a942a7f320ed96230ece4222829fa584b58abe4ffcfa96ebb"} Dec 01 08:56:04 crc kubenswrapper[4873]: I1201 08:56:04.989370 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-bhrpn" event={"ID":"225323c8-ed2f-4573-920d-f43f3e4561de","Type":"ContainerStarted","Data":"61417d74aad8ed143200738ab3a884ddf394649a80c6419ccb9ce95c8aad5745"} Dec 01 08:56:04 crc kubenswrapper[4873]: I1201 08:56:04.997787 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-gqm5w" podStartSLOduration=11.950618338 podStartE2EDuration="40.997756748s" podCreationTimestamp="2025-12-01 08:55:24 +0000 UTC" firstStartedPulling="2025-12-01 08:55:26.422723881 +0000 UTC m=+902.324832420" lastFinishedPulling="2025-12-01 08:55:55.469862291 +0000 UTC m=+931.371970830" observedRunningTime="2025-12-01 08:56:04.988801345 +0000 UTC m=+940.890909894" watchObservedRunningTime="2025-12-01 08:56:04.997756748 +0000 UTC m=+940.899865287" Dec 01 08:56:04 crc kubenswrapper[4873]: I1201 08:56:04.999372 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-n79q9" event={"ID":"7e3a2af8-3381-46e2-8c23-41aab8fd1a5e","Type":"ContainerStarted","Data":"0d45285d2b1927269bd94a830dd968d3949df44e410f727c86d7747cc9090999"} Dec 01 08:56:05 crc kubenswrapper[4873]: I1201 08:56:05.008582 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-45whz" event={"ID":"94e7345b-b1ac-46e2-be25-9d64d3d33523","Type":"ContainerStarted","Data":"f2ea67f1100ea3ce20923c8a09d7f29917b0b2bd3951990d57a8e28e02e0f185"} Dec 01 08:56:05 crc kubenswrapper[4873]: I1201 08:56:05.028438 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-49bz9" event={"ID":"879e02ea-306e-4e7b-9012-cb1cb6bdee00","Type":"ContainerStarted","Data":"910533b8c04caef8be41f549f9205abc40e06150112bd975b4e24743c7b19f65"} Dec 01 08:56:05 crc kubenswrapper[4873]: I1201 08:56:05.036886 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-d7c6w" event={"ID":"967215e8-7a18-4507-96c1-4c79c7e1d51a","Type":"ContainerStarted","Data":"7e18bb0bc0ffdba1b75f3fb5b41beaec51d156f7e2701f7512c318cc906ca5f3"} Dec 01 08:56:05 crc kubenswrapper[4873]: I1201 08:56:05.036949 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-d7c6w" Dec 01 08:56:05 crc kubenswrapper[4873]: E1201 08:56:05.054888 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670\\\"\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-x9kc4" podUID="450a3417-0074-4223-b1e3-aa1b854320fe" Dec 01 08:56:05 crc kubenswrapper[4873]: I1201 08:56:05.135077 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-d7c6w" podStartSLOduration=13.760751496 podStartE2EDuration="41.135054024s" podCreationTimestamp="2025-12-01 08:55:24 +0000 UTC" firstStartedPulling="2025-12-01 08:55:28.09662008 +0000 UTC m=+903.998728619" lastFinishedPulling="2025-12-01 08:55:55.470922598 +0000 UTC m=+931.373031147" observedRunningTime="2025-12-01 08:56:05.133903575 +0000 UTC m=+941.036012114" watchObservedRunningTime="2025-12-01 08:56:05.135054024 +0000 UTC m=+941.037162563" Dec 01 08:56:05 crc kubenswrapper[4873]: E1201 08:56:05.150957 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-pfphw" podUID="035095bd-1b65-4895-95a1-59feee524920" Dec 01 08:56:05 crc kubenswrapper[4873]: E1201 08:56:05.188169 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-dx4jm" podUID="1cd3aa85-6333-4cfd-aacf-e51169ef4b42" Dec 01 08:56:05 crc kubenswrapper[4873]: I1201 08:56:05.299413 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-49bz9" podStartSLOduration=8.10326434 podStartE2EDuration="41.299394212s" podCreationTimestamp="2025-12-01 08:55:24 +0000 UTC" firstStartedPulling="2025-12-01 08:55:28.812852147 +0000 UTC m=+904.714960676" lastFinishedPulling="2025-12-01 08:56:02.008982009 +0000 UTC m=+937.911090548" observedRunningTime="2025-12-01 08:56:05.297780712 +0000 UTC m=+941.199889251" watchObservedRunningTime="2025-12-01 08:56:05.299394212 +0000 UTC m=+941.201502751" Dec 01 08:56:05 crc kubenswrapper[4873]: E1201 08:56:05.308498 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5rq8z" podUID="51253ba6-e0b5-44ac-8c18-be17b4d13024" Dec 01 08:56:05 crc kubenswrapper[4873]: E1201 08:56:05.356727 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7wmqg" podUID="eb53d542-63e9-487d-9d06-237c4b2b9252" Dec 01 08:56:05 crc kubenswrapper[4873]: E1201 08:56:05.357041 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-hnzzf" podUID="3eff7ce2-9ff7-413c-b472-9e114e7130ca" Dec 01 08:56:05 crc kubenswrapper[4873]: I1201 08:56:05.663548 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-49bz9" Dec 01 08:56:06 crc kubenswrapper[4873]: I1201 08:56:06.087543 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-bhrpn" event={"ID":"225323c8-ed2f-4573-920d-f43f3e4561de","Type":"ContainerStarted","Data":"d610a51e82b472ce303a91cce0828f40b27e74434c960ffed0d4b47a182a6b35"} Dec 01 08:56:06 crc kubenswrapper[4873]: I1201 08:56:06.108080 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z2r2m" event={"ID":"f8f3966d-9400-44c1-a3e0-f4236541480e","Type":"ContainerStarted","Data":"978cabc5f4a7ff33f74ff41b164eb6ab71d1f0a87af985872823c682b3c75bcf"} Dec 01 08:56:06 crc kubenswrapper[4873]: I1201 08:56:06.129850 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7f4d6cf897-fddkc" event={"ID":"223a64ee-92cb-4b29-91bf-ffa7ed7d64ce","Type":"ContainerStarted","Data":"d3941194c6a012e0ab7e5d1b258e8504d1a6555ef894cc5e11b175b4f68ad27d"} Dec 01 08:56:06 crc kubenswrapper[4873]: I1201 08:56:06.130046 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-7f4d6cf897-fddkc" Dec 01 08:56:06 crc kubenswrapper[4873]: I1201 08:56:06.154973 4873 generic.go:334] "Generic (PLEG): container finished" podID="074c5f85-694d-47af-a33f-e1995ae73a5c" containerID="ae86de4f1adff7a38a04222ea6228b7ecf71a389fd4a05dfe9c5150cc951b5b5" exitCode=0 Dec 01 08:56:06 crc kubenswrapper[4873]: I1201 08:56:06.155307 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f6n4k" event={"ID":"074c5f85-694d-47af-a33f-e1995ae73a5c","Type":"ContainerDied","Data":"ae86de4f1adff7a38a04222ea6228b7ecf71a389fd4a05dfe9c5150cc951b5b5"} Dec 01 08:56:06 crc kubenswrapper[4873]: I1201 08:56:06.190478 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-v7bvz" event={"ID":"3e0ad069-d971-4fd0-93df-b6a8794afe00","Type":"ContainerStarted","Data":"340c6f5a6ad714b1ccfa6a6af08654526e35705e154dc66f3f483b7921ebdb78"} Dec 01 08:56:06 crc kubenswrapper[4873]: I1201 08:56:06.191848 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-v7bvz" Dec 01 08:56:06 crc kubenswrapper[4873]: I1201 08:56:06.203511 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-7f4d6cf897-fddkc" podStartSLOduration=42.203475512 podStartE2EDuration="42.203475512s" podCreationTimestamp="2025-12-01 08:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:56:06.190769516 +0000 UTC m=+942.092878055" watchObservedRunningTime="2025-12-01 08:56:06.203475512 +0000 UTC m=+942.105584051" Dec 01 08:56:06 crc kubenswrapper[4873]: I1201 08:56:06.223835 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5cd474b4ff-wcmv4" event={"ID":"f778fd6c-4c8a-4067-b3f2-cb7d98a50bf7","Type":"ContainerStarted","Data":"eb8988abb727c944c216737f60b3bb4a0328a8be59ff909929ec362f1cd87a4e"} Dec 01 08:56:06 crc kubenswrapper[4873]: I1201 08:56:06.224313 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-5cd474b4ff-wcmv4" Dec 01 08:56:06 crc kubenswrapper[4873]: I1201 08:56:06.256611 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-hnzzf" event={"ID":"3eff7ce2-9ff7-413c-b472-9e114e7130ca","Type":"ContainerStarted","Data":"1bf235a6bed01d63cb3c79559e39c90c947b1fd1764de56e5d3946614657b1fc"} Dec 01 08:56:06 crc kubenswrapper[4873]: I1201 08:56:06.264575 4873 generic.go:334] "Generic (PLEG): container finished" podID="77cdace7-aa56-4a6b-900b-b162ef0410a4" containerID="13dd90155611dfdb26ad3895edb8ecab8b2a44a0d59464b7ffc4eda7ac39b901" exitCode=0 Dec 01 08:56:06 crc kubenswrapper[4873]: I1201 08:56:06.264669 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tdc2x" event={"ID":"77cdace7-aa56-4a6b-900b-b162ef0410a4","Type":"ContainerDied","Data":"13dd90155611dfdb26ad3895edb8ecab8b2a44a0d59464b7ffc4eda7ac39b901"} Dec 01 08:56:06 crc kubenswrapper[4873]: I1201 08:56:06.277318 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-v7bvz" podStartSLOduration=9.048858584 podStartE2EDuration="42.277293029s" podCreationTimestamp="2025-12-01 08:55:24 +0000 UTC" firstStartedPulling="2025-12-01 08:55:28.789302392 +0000 UTC m=+904.691410931" lastFinishedPulling="2025-12-01 08:56:02.017736837 +0000 UTC m=+937.919845376" observedRunningTime="2025-12-01 08:56:06.274447378 +0000 UTC m=+942.176555917" watchObservedRunningTime="2025-12-01 08:56:06.277293029 +0000 UTC m=+942.179401568" Dec 01 08:56:06 crc kubenswrapper[4873]: I1201 08:56:06.294493 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-dx4jm" event={"ID":"1cd3aa85-6333-4cfd-aacf-e51169ef4b42","Type":"ContainerStarted","Data":"b456d5c7bff462b80d45698a680b8d93ec884d8def4541040fd2a76d7f1cc811"} Dec 01 08:56:06 crc kubenswrapper[4873]: I1201 08:56:06.313252 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5rq8z" event={"ID":"51253ba6-e0b5-44ac-8c18-be17b4d13024","Type":"ContainerStarted","Data":"7d582a31000b7d676dc00b34f8ee9335fb9c290d870799670ee01de4d1fa6b3f"} Dec 01 08:56:06 crc kubenswrapper[4873]: I1201 08:56:06.349874 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-5cd474b4ff-wcmv4" podStartSLOduration=4.808811177 podStartE2EDuration="42.349842904s" podCreationTimestamp="2025-12-01 08:55:24 +0000 UTC" firstStartedPulling="2025-12-01 08:55:27.945238614 +0000 UTC m=+903.847347153" lastFinishedPulling="2025-12-01 08:56:05.486270341 +0000 UTC m=+941.388378880" observedRunningTime="2025-12-01 08:56:06.346918301 +0000 UTC m=+942.249026840" watchObservedRunningTime="2025-12-01 08:56:06.349842904 +0000 UTC m=+942.251951443" Dec 01 08:56:06 crc kubenswrapper[4873]: I1201 08:56:06.355571 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-pfphw" event={"ID":"035095bd-1b65-4895-95a1-59feee524920","Type":"ContainerStarted","Data":"9fed6f26eaa5fc3801f2e484a94aa67cea7ceae98121105fde7f0d8f84259e5e"} Dec 01 08:56:06 crc kubenswrapper[4873]: E1201 08:56:06.364362 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:986861e5a0a9954f63581d9d55a30f8057883cefea489415d76257774526eea3\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-pfphw" podUID="035095bd-1b65-4895-95a1-59feee524920" Dec 01 08:56:06 crc kubenswrapper[4873]: I1201 08:56:06.376736 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-mvc26" event={"ID":"34e3726b-c9f5-4a86-9e87-e179f8dae739","Type":"ContainerStarted","Data":"2b7ac79673f5c911a8b4530bcde840b7d1d601c81af0ffb187507ac4cdabc07a"} Dec 01 08:56:06 crc kubenswrapper[4873]: I1201 08:56:06.378285 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-mvc26" Dec 01 08:56:06 crc kubenswrapper[4873]: I1201 08:56:06.400446 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-j2vnt" event={"ID":"b45909ce-2ea4-4cf2-9351-c8839e44d734","Type":"ContainerStarted","Data":"fe894a621baeaec7408cbba3610c15ef903e8a596e55d401d4a4cde57549e78e"} Dec 01 08:56:06 crc kubenswrapper[4873]: I1201 08:56:06.401596 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-j2vnt" Dec 01 08:56:06 crc kubenswrapper[4873]: I1201 08:56:06.406271 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-sjw48" event={"ID":"5ec1188a-c7f9-4cc8-8ea9-c1f1977041a5","Type":"ContainerStarted","Data":"56be1a1343c45af2b781868ac5532e8c1be7dde814b77c19fc9b2a8c16cfcfc1"} Dec 01 08:56:06 crc kubenswrapper[4873]: I1201 08:56:06.407415 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-sjw48" Dec 01 08:56:06 crc kubenswrapper[4873]: I1201 08:56:06.441324 4873 generic.go:334] "Generic (PLEG): container finished" podID="439ab38e-6a5b-4aff-baef-0be711a2fad5" containerID="a9c04480d5d9be7bc887ae98d780d9b51325cf96f9a0c3806a4a1d9156bbaafa" exitCode=0 Dec 01 08:56:06 crc kubenswrapper[4873]: I1201 08:56:06.589888 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jvrlm" event={"ID":"439ab38e-6a5b-4aff-baef-0be711a2fad5","Type":"ContainerDied","Data":"a9c04480d5d9be7bc887ae98d780d9b51325cf96f9a0c3806a4a1d9156bbaafa"} Dec 01 08:56:06 crc kubenswrapper[4873]: I1201 08:56:06.590323 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7wmqg" event={"ID":"eb53d542-63e9-487d-9d06-237c4b2b9252","Type":"ContainerStarted","Data":"cffbc8a5daba2610ae49f3771dd16a4265ae8e4d48537307d4cc9fa225aaa833"} Dec 01 08:56:06 crc kubenswrapper[4873]: I1201 08:56:06.598473 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-6r8hb" event={"ID":"723f446d-21bb-432a-b2af-93219593819d","Type":"ContainerStarted","Data":"ecff933480f1e1fd4d2fad355f64cb43e9bf50995d58c4e431e968cf5078a18f"} Dec 01 08:56:06 crc kubenswrapper[4873]: I1201 08:56:06.598696 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-6r8hb" Dec 01 08:56:06 crc kubenswrapper[4873]: I1201 08:56:06.888340 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-mvc26" podStartSLOduration=15.813333076 podStartE2EDuration="42.888313358s" podCreationTimestamp="2025-12-01 08:55:24 +0000 UTC" firstStartedPulling="2025-12-01 08:55:28.395943756 +0000 UTC m=+904.298052295" lastFinishedPulling="2025-12-01 08:55:55.470924038 +0000 UTC m=+931.373032577" observedRunningTime="2025-12-01 08:56:06.572234496 +0000 UTC m=+942.474343035" watchObservedRunningTime="2025-12-01 08:56:06.888313358 +0000 UTC m=+942.790421897" Dec 01 08:56:06 crc kubenswrapper[4873]: I1201 08:56:06.931845 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-sjw48" podStartSLOduration=15.513370045 podStartE2EDuration="42.93181581s" podCreationTimestamp="2025-12-01 08:55:24 +0000 UTC" firstStartedPulling="2025-12-01 08:55:28.052503813 +0000 UTC m=+903.954612352" lastFinishedPulling="2025-12-01 08:55:55.470949578 +0000 UTC m=+931.373058117" observedRunningTime="2025-12-01 08:56:06.929518113 +0000 UTC m=+942.831626652" watchObservedRunningTime="2025-12-01 08:56:06.93181581 +0000 UTC m=+942.833924349" Dec 01 08:56:07 crc kubenswrapper[4873]: I1201 08:56:07.087221 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-j2vnt" podStartSLOduration=6.228747568 podStartE2EDuration="43.087200325s" podCreationTimestamp="2025-12-01 08:55:24 +0000 UTC" firstStartedPulling="2025-12-01 08:55:28.263286976 +0000 UTC m=+904.165395515" lastFinishedPulling="2025-12-01 08:56:05.121739733 +0000 UTC m=+941.023848272" observedRunningTime="2025-12-01 08:56:07.085802701 +0000 UTC m=+942.987911240" watchObservedRunningTime="2025-12-01 08:56:07.087200325 +0000 UTC m=+942.989308864" Dec 01 08:56:07 crc kubenswrapper[4873]: I1201 08:56:07.095628 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-6r8hb" podStartSLOduration=14.425832042 podStartE2EDuration="43.095613805s" podCreationTimestamp="2025-12-01 08:55:24 +0000 UTC" firstStartedPulling="2025-12-01 08:55:26.801400481 +0000 UTC m=+902.703509020" lastFinishedPulling="2025-12-01 08:55:55.471182244 +0000 UTC m=+931.373290783" observedRunningTime="2025-12-01 08:56:07.051480367 +0000 UTC m=+942.953588926" watchObservedRunningTime="2025-12-01 08:56:07.095613805 +0000 UTC m=+942.997722344" Dec 01 08:56:07 crc kubenswrapper[4873]: I1201 08:56:07.628147 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-45whz" event={"ID":"94e7345b-b1ac-46e2-be25-9d64d3d33523","Type":"ContainerStarted","Data":"cac036ee692f55eaa25d19b839c159bf5f60d44955aaa3917187a8a1c2c0eb0b"} Dec 01 08:56:07 crc kubenswrapper[4873]: I1201 08:56:07.664923 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-45whz" Dec 01 08:56:07 crc kubenswrapper[4873]: I1201 08:56:07.665050 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-6r8hb" Dec 01 08:56:07 crc kubenswrapper[4873]: I1201 08:56:07.665092 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-sjw48" Dec 01 08:56:07 crc kubenswrapper[4873]: I1201 08:56:07.671558 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-49bz9" Dec 01 08:56:07 crc kubenswrapper[4873]: I1201 08:56:07.677609 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-mvc26" Dec 01 08:56:07 crc kubenswrapper[4873]: I1201 08:56:07.714689 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-45whz" podStartSLOduration=6.126616349 podStartE2EDuration="43.714661835s" podCreationTimestamp="2025-12-01 08:55:24 +0000 UTC" firstStartedPulling="2025-12-01 08:55:28.31329677 +0000 UTC m=+904.215405309" lastFinishedPulling="2025-12-01 08:56:05.901342256 +0000 UTC m=+941.803450795" observedRunningTime="2025-12-01 08:56:07.702461541 +0000 UTC m=+943.604570100" watchObservedRunningTime="2025-12-01 08:56:07.714661835 +0000 UTC m=+943.616770374" Dec 01 08:56:08 crc kubenswrapper[4873]: I1201 08:56:08.071297 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-bhrpn" podStartSLOduration=6.731079156 podStartE2EDuration="44.071278326s" podCreationTimestamp="2025-12-01 08:55:24 +0000 UTC" firstStartedPulling="2025-12-01 08:55:28.388092071 +0000 UTC m=+904.290200610" lastFinishedPulling="2025-12-01 08:56:05.728291241 +0000 UTC m=+941.630399780" observedRunningTime="2025-12-01 08:56:08.066311282 +0000 UTC m=+943.968419821" watchObservedRunningTime="2025-12-01 08:56:08.071278326 +0000 UTC m=+943.973386865" Dec 01 08:56:08 crc kubenswrapper[4873]: I1201 08:56:08.675828 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-bhrpn" Dec 01 08:56:08 crc kubenswrapper[4873]: I1201 08:56:08.695668 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-v7bvz" Dec 01 08:56:09 crc kubenswrapper[4873]: I1201 08:56:09.754385 4873 generic.go:334] "Generic (PLEG): container finished" podID="f8f3966d-9400-44c1-a3e0-f4236541480e" containerID="978cabc5f4a7ff33f74ff41b164eb6ab71d1f0a87af985872823c682b3c75bcf" exitCode=0 Dec 01 08:56:09 crc kubenswrapper[4873]: I1201 08:56:09.754511 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z2r2m" event={"ID":"f8f3966d-9400-44c1-a3e0-f4236541480e","Type":"ContainerDied","Data":"978cabc5f4a7ff33f74ff41b164eb6ab71d1f0a87af985872823c682b3c75bcf"} Dec 01 08:56:11 crc kubenswrapper[4873]: I1201 08:56:11.570779 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-7f4d6cf897-fddkc" Dec 01 08:56:14 crc kubenswrapper[4873]: I1201 08:56:14.442729 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-gqm5w" Dec 01 08:56:14 crc kubenswrapper[4873]: I1201 08:56:14.589432 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-d7c6w" Dec 01 08:56:14 crc kubenswrapper[4873]: I1201 08:56:14.714273 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-45whz" Dec 01 08:56:14 crc kubenswrapper[4873]: I1201 08:56:14.971161 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-5cd474b4ff-wcmv4" Dec 01 08:56:15 crc kubenswrapper[4873]: I1201 08:56:15.611901 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-j2vnt" Dec 01 08:56:15 crc kubenswrapper[4873]: I1201 08:56:15.875351 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-bhrpn" Dec 01 08:56:19 crc kubenswrapper[4873]: E1201 08:56:19.326102 4873 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/infra-operator@sha256:09a6d0613ee2d3c1c809fc36c22678458ac271e0da87c970aec0a5339f5423f7" Dec 01 08:56:19 crc kubenswrapper[4873]: E1201 08:56:19.326816 4873 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:09a6d0613ee2d3c1c809fc36c22678458ac271e0da87c970aec0a5339f5423f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9klvb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-57548d458d-tmdj7_openstack-operators(4dcc939e-b548-4fb6-814a-30e2aaa8a94a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 08:56:19 crc kubenswrapper[4873]: I1201 08:56:19.431165 4873 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 08:56:20 crc kubenswrapper[4873]: E1201 08:56:20.271987 4873 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:14cfad6ea2e7f7ecc4cb2aafceb9c61514b3d04b66668832d1e4ac3b19f1ab81" Dec 01 08:56:20 crc kubenswrapper[4873]: E1201 08:56:20.272589 4873 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:14cfad6ea2e7f7ecc4cb2aafceb9c61514b3d04b66668832d1e4ac3b19f1ab81,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter:v0.15.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_API_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_PROC_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-processor:current,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler:release-0.7.12,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter:v1.5.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter:v1.10.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine:current-podified,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xc57g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-64bc77cfd4vf8qt_openstack-operators(5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 08:56:20 crc kubenswrapper[4873]: E1201 08:56:20.811721 4873 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d" Dec 01 08:56:20 crc kubenswrapper[4873]: E1201 08:56:20.811954 4873 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nrqtf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-6dxcw_openstack-operators(f7a7bcca-9403-4e2d-bc28-53d4eb5ae252): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 08:56:21 crc kubenswrapper[4873]: E1201 08:56:21.545285 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vf8qt" podUID="5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5" Dec 01 08:56:21 crc kubenswrapper[4873]: E1201 08:56:21.609554 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6dxcw" podUID="f7a7bcca-9403-4e2d-bc28-53d4eb5ae252" Dec 01 08:56:21 crc kubenswrapper[4873]: E1201 08:56:21.647927 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/infra-operator-controller-manager-57548d458d-tmdj7" podUID="4dcc939e-b548-4fb6-814a-30e2aaa8a94a" Dec 01 08:56:21 crc kubenswrapper[4873]: I1201 08:56:21.921187 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-tmdj7" event={"ID":"4dcc939e-b548-4fb6-814a-30e2aaa8a94a","Type":"ContainerStarted","Data":"d39f6a18ee8f4575acc11242d29fc84f64901399452b85267f899aa0d66b86fa"} Dec 01 08:56:21 crc kubenswrapper[4873]: E1201 08:56:21.924170 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:09a6d0613ee2d3c1c809fc36c22678458ac271e0da87c970aec0a5339f5423f7\\\"\"" pod="openstack-operators/infra-operator-controller-manager-57548d458d-tmdj7" podUID="4dcc939e-b548-4fb6-814a-30e2aaa8a94a" Dec 01 08:56:21 crc kubenswrapper[4873]: I1201 08:56:21.930063 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-x9kc4" event={"ID":"450a3417-0074-4223-b1e3-aa1b854320fe","Type":"ContainerStarted","Data":"43a4a1491320d6999d81ebedcf685c8d0d8153f4416dd21a7598c2e49982c98a"} Dec 01 08:56:21 crc kubenswrapper[4873]: I1201 08:56:21.931058 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-x9kc4" Dec 01 08:56:21 crc kubenswrapper[4873]: I1201 08:56:21.932668 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vf8qt" event={"ID":"5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5","Type":"ContainerStarted","Data":"c8c983d857bda156e6e0a14b096841ce0037e608fe9078f39413cd457ba899fa"} Dec 01 08:56:21 crc kubenswrapper[4873]: E1201 08:56:21.934050 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:14cfad6ea2e7f7ecc4cb2aafceb9c61514b3d04b66668832d1e4ac3b19f1ab81\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vf8qt" podUID="5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5" Dec 01 08:56:21 crc kubenswrapper[4873]: I1201 08:56:21.935593 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-pfphw" event={"ID":"035095bd-1b65-4895-95a1-59feee524920","Type":"ContainerStarted","Data":"93836dbbe0f0a85566e48b850cdc40482b21077c949ed3e4604d51d2cf636565"} Dec 01 08:56:21 crc kubenswrapper[4873]: I1201 08:56:21.936305 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-pfphw" Dec 01 08:56:21 crc kubenswrapper[4873]: I1201 08:56:21.937954 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-dx4jm" event={"ID":"1cd3aa85-6333-4cfd-aacf-e51169ef4b42","Type":"ContainerStarted","Data":"314d6ed49adca9bbd9c1d8208bc075e754fc71583abcbc96f7460afe15d5721a"} Dec 01 08:56:21 crc kubenswrapper[4873]: I1201 08:56:21.938441 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-dx4jm" Dec 01 08:56:21 crc kubenswrapper[4873]: I1201 08:56:21.939800 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-hnzzf" event={"ID":"3eff7ce2-9ff7-413c-b472-9e114e7130ca","Type":"ContainerStarted","Data":"e7c491f4d63dcdfe5d3718f702d4741a58d55eb2b9f123963cd845a2616bab55"} Dec 01 08:56:21 crc kubenswrapper[4873]: I1201 08:56:21.940284 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-hnzzf" Dec 01 08:56:21 crc kubenswrapper[4873]: I1201 08:56:21.941727 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5rq8z" event={"ID":"51253ba6-e0b5-44ac-8c18-be17b4d13024","Type":"ContainerStarted","Data":"ed90d08ea64bd4cf4ff64ad2f035284e0e913ed3fd1c00ec0f59c295798fafa7"} Dec 01 08:56:21 crc kubenswrapper[4873]: I1201 08:56:21.942229 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5rq8z" Dec 01 08:56:21 crc kubenswrapper[4873]: I1201 08:56:21.943666 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7wmqg" event={"ID":"eb53d542-63e9-487d-9d06-237c4b2b9252","Type":"ContainerStarted","Data":"3384213b0ce2a927183b1f9f754ac4a46379a19069fb38f1e6bd77d3479a14d6"} Dec 01 08:56:21 crc kubenswrapper[4873]: I1201 08:56:21.944167 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7wmqg" Dec 01 08:56:21 crc kubenswrapper[4873]: I1201 08:56:21.945710 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jvrlm" event={"ID":"439ab38e-6a5b-4aff-baef-0be711a2fad5","Type":"ContainerStarted","Data":"ad585e8259e362201e754008e702c73364e7a1a57452d275026cb18e63c71963"} Dec 01 08:56:21 crc kubenswrapper[4873]: I1201 08:56:21.947470 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-n79q9" event={"ID":"7e3a2af8-3381-46e2-8c23-41aab8fd1a5e","Type":"ContainerStarted","Data":"de3e5b52f6fbb60ff0341dc9dc6c5a8adfeeee5dd8884cefb2bed8cffecd2d24"} Dec 01 08:56:21 crc kubenswrapper[4873]: I1201 08:56:21.947931 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-n79q9" Dec 01 08:56:21 crc kubenswrapper[4873]: I1201 08:56:21.949597 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z2r2m" event={"ID":"f8f3966d-9400-44c1-a3e0-f4236541480e","Type":"ContainerStarted","Data":"42705b68e74d03f05c301a50433983cd22d82d47e640c48090dbef3efb037824"} Dec 01 08:56:21 crc kubenswrapper[4873]: I1201 08:56:21.951370 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6dxcw" event={"ID":"f7a7bcca-9403-4e2d-bc28-53d4eb5ae252","Type":"ContainerStarted","Data":"555b15be1d35f444dcb330fa2596f8c5e0c61cae7f21ebf1b68c702925eac7cc"} Dec 01 08:56:21 crc kubenswrapper[4873]: E1201 08:56:21.952530 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\"" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6dxcw" podUID="f7a7bcca-9403-4e2d-bc28-53d4eb5ae252" Dec 01 08:56:21 crc kubenswrapper[4873]: I1201 08:56:21.954461 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tdc2x" event={"ID":"77cdace7-aa56-4a6b-900b-b162ef0410a4","Type":"ContainerStarted","Data":"f5daf524faa67d92cf347a0d236cacad6068db9373ed4987baa86d3d89d5f636"} Dec 01 08:56:21 crc kubenswrapper[4873]: I1201 08:56:21.956642 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f6n4k" event={"ID":"074c5f85-694d-47af-a33f-e1995ae73a5c","Type":"ContainerStarted","Data":"d6b1c3d0d1b343c324e7107882e34aeced110702cd6413eb2a36c9606e32d77b"} Dec 01 08:56:22 crc kubenswrapper[4873]: I1201 08:56:22.395666 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7wmqg" podStartSLOduration=5.53419078 podStartE2EDuration="58.39564462s" podCreationTimestamp="2025-12-01 08:55:24 +0000 UTC" firstStartedPulling="2025-12-01 08:55:27.953457529 +0000 UTC m=+903.855566068" lastFinishedPulling="2025-12-01 08:56:20.814911369 +0000 UTC m=+956.717019908" observedRunningTime="2025-12-01 08:56:22.394149713 +0000 UTC m=+958.296258252" watchObservedRunningTime="2025-12-01 08:56:22.39564462 +0000 UTC m=+958.297753159" Dec 01 08:56:22 crc kubenswrapper[4873]: I1201 08:56:22.475649 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5rq8z" podStartSLOduration=3.928687633 podStartE2EDuration="58.47561741s" podCreationTimestamp="2025-12-01 08:55:24 +0000 UTC" firstStartedPulling="2025-12-01 08:55:26.32218928 +0000 UTC m=+902.224297819" lastFinishedPulling="2025-12-01 08:56:20.869119057 +0000 UTC m=+956.771227596" observedRunningTime="2025-12-01 08:56:22.469498868 +0000 UTC m=+958.371607407" watchObservedRunningTime="2025-12-01 08:56:22.47561741 +0000 UTC m=+958.377725959" Dec 01 08:56:22 crc kubenswrapper[4873]: I1201 08:56:22.533890 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-hnzzf" podStartSLOduration=6.039393879 podStartE2EDuration="58.533868729s" podCreationTimestamp="2025-12-01 08:55:24 +0000 UTC" firstStartedPulling="2025-12-01 08:55:28.381007515 +0000 UTC m=+904.283116054" lastFinishedPulling="2025-12-01 08:56:20.875482365 +0000 UTC m=+956.777590904" observedRunningTime="2025-12-01 08:56:22.530407083 +0000 UTC m=+958.432515622" watchObservedRunningTime="2025-12-01 08:56:22.533868729 +0000 UTC m=+958.435977268" Dec 01 08:56:22 crc kubenswrapper[4873]: I1201 08:56:22.536583 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-x9kc4" podStartSLOduration=5.678304406 podStartE2EDuration="58.536566126s" podCreationTimestamp="2025-12-01 08:55:24 +0000 UTC" firstStartedPulling="2025-12-01 08:55:28.034250409 +0000 UTC m=+903.936358948" lastFinishedPulling="2025-12-01 08:56:20.892512129 +0000 UTC m=+956.794620668" observedRunningTime="2025-12-01 08:56:22.494790257 +0000 UTC m=+958.396898796" watchObservedRunningTime="2025-12-01 08:56:22.536566126 +0000 UTC m=+958.438674665" Dec 01 08:56:22 crc kubenswrapper[4873]: I1201 08:56:22.574324 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-pfphw" podStartSLOduration=6.316216495 podStartE2EDuration="58.574302565s" podCreationTimestamp="2025-12-01 08:55:24 +0000 UTC" firstStartedPulling="2025-12-01 08:55:28.557168417 +0000 UTC m=+904.459276956" lastFinishedPulling="2025-12-01 08:56:20.815254497 +0000 UTC m=+956.717363026" observedRunningTime="2025-12-01 08:56:22.572632703 +0000 UTC m=+958.474741242" watchObservedRunningTime="2025-12-01 08:56:22.574302565 +0000 UTC m=+958.476411104" Dec 01 08:56:22 crc kubenswrapper[4873]: I1201 08:56:22.720516 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-dx4jm" podStartSLOduration=6.475475777 podStartE2EDuration="58.720489742s" podCreationTimestamp="2025-12-01 08:55:24 +0000 UTC" firstStartedPulling="2025-12-01 08:55:28.619149919 +0000 UTC m=+904.521258458" lastFinishedPulling="2025-12-01 08:56:20.864163884 +0000 UTC m=+956.766272423" observedRunningTime="2025-12-01 08:56:22.618353261 +0000 UTC m=+958.520461810" watchObservedRunningTime="2025-12-01 08:56:22.720489742 +0000 UTC m=+958.622598281" Dec 01 08:56:22 crc kubenswrapper[4873]: I1201 08:56:22.722941 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-z2r2m" podStartSLOduration=23.76644508 podStartE2EDuration="40.722935783s" podCreationTimestamp="2025-12-01 08:55:42 +0000 UTC" firstStartedPulling="2025-12-01 08:56:03.93256297 +0000 UTC m=+939.834671509" lastFinishedPulling="2025-12-01 08:56:20.889053663 +0000 UTC m=+956.791162212" observedRunningTime="2025-12-01 08:56:22.717777054 +0000 UTC m=+958.619885593" watchObservedRunningTime="2025-12-01 08:56:22.722935783 +0000 UTC m=+958.625044322" Dec 01 08:56:22 crc kubenswrapper[4873]: I1201 08:56:22.760504 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-n79q9" podStartSLOduration=6.224287638 podStartE2EDuration="58.760477716s" podCreationTimestamp="2025-12-01 08:55:24 +0000 UTC" firstStartedPulling="2025-12-01 08:55:28.339479412 +0000 UTC m=+904.241587941" lastFinishedPulling="2025-12-01 08:56:20.87566948 +0000 UTC m=+956.777778019" observedRunningTime="2025-12-01 08:56:22.752264012 +0000 UTC m=+958.654372551" watchObservedRunningTime="2025-12-01 08:56:22.760477716 +0000 UTC m=+958.662586255" Dec 01 08:56:22 crc kubenswrapper[4873]: E1201 08:56:22.969551 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:14cfad6ea2e7f7ecc4cb2aafceb9c61514b3d04b66668832d1e4ac3b19f1ab81\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vf8qt" podUID="5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5" Dec 01 08:56:22 crc kubenswrapper[4873]: E1201 08:56:22.969861 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\"" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6dxcw" podUID="f7a7bcca-9403-4e2d-bc28-53d4eb5ae252" Dec 01 08:56:22 crc kubenswrapper[4873]: E1201 08:56:22.969937 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:09a6d0613ee2d3c1c809fc36c22678458ac271e0da87c970aec0a5339f5423f7\\\"\"" pod="openstack-operators/infra-operator-controller-manager-57548d458d-tmdj7" podUID="4dcc939e-b548-4fb6-814a-30e2aaa8a94a" Dec 01 08:56:23 crc kubenswrapper[4873]: I1201 08:56:23.033812 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-z2r2m" Dec 01 08:56:23 crc kubenswrapper[4873]: I1201 08:56:23.033871 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-z2r2m" Dec 01 08:56:24 crc kubenswrapper[4873]: I1201 08:56:24.003537 4873 generic.go:334] "Generic (PLEG): container finished" podID="77cdace7-aa56-4a6b-900b-b162ef0410a4" containerID="f5daf524faa67d92cf347a0d236cacad6068db9373ed4987baa86d3d89d5f636" exitCode=0 Dec 01 08:56:24 crc kubenswrapper[4873]: I1201 08:56:24.003640 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tdc2x" event={"ID":"77cdace7-aa56-4a6b-900b-b162ef0410a4","Type":"ContainerDied","Data":"f5daf524faa67d92cf347a0d236cacad6068db9373ed4987baa86d3d89d5f636"} Dec 01 08:56:24 crc kubenswrapper[4873]: I1201 08:56:24.005890 4873 generic.go:334] "Generic (PLEG): container finished" podID="074c5f85-694d-47af-a33f-e1995ae73a5c" containerID="d6b1c3d0d1b343c324e7107882e34aeced110702cd6413eb2a36c9606e32d77b" exitCode=0 Dec 01 08:56:24 crc kubenswrapper[4873]: I1201 08:56:24.005971 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f6n4k" event={"ID":"074c5f85-694d-47af-a33f-e1995ae73a5c","Type":"ContainerDied","Data":"d6b1c3d0d1b343c324e7107882e34aeced110702cd6413eb2a36c9606e32d77b"} Dec 01 08:56:24 crc kubenswrapper[4873]: I1201 08:56:24.008440 4873 generic.go:334] "Generic (PLEG): container finished" podID="439ab38e-6a5b-4aff-baef-0be711a2fad5" containerID="ad585e8259e362201e754008e702c73364e7a1a57452d275026cb18e63c71963" exitCode=0 Dec 01 08:56:24 crc kubenswrapper[4873]: I1201 08:56:24.008488 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jvrlm" event={"ID":"439ab38e-6a5b-4aff-baef-0be711a2fad5","Type":"ContainerDied","Data":"ad585e8259e362201e754008e702c73364e7a1a57452d275026cb18e63c71963"} Dec 01 08:56:24 crc kubenswrapper[4873]: I1201 08:56:24.273100 4873 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-z2r2m" podUID="f8f3966d-9400-44c1-a3e0-f4236541480e" containerName="registry-server" probeResult="failure" output=< Dec 01 08:56:24 crc kubenswrapper[4873]: timeout: failed to connect service ":50051" within 1s Dec 01 08:56:24 crc kubenswrapper[4873]: > Dec 01 08:56:25 crc kubenswrapper[4873]: I1201 08:56:25.018713 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tdc2x" event={"ID":"77cdace7-aa56-4a6b-900b-b162ef0410a4","Type":"ContainerStarted","Data":"a70421f9712cd45f73b9a9ff27639f982bfbe63db77ce7c7f4477b42979c2d27"} Dec 01 08:56:25 crc kubenswrapper[4873]: I1201 08:56:25.022292 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f6n4k" event={"ID":"074c5f85-694d-47af-a33f-e1995ae73a5c","Type":"ContainerStarted","Data":"266158624fe4045657a423d2622eeda0d8688b0f02b87848d975907c902d88fb"} Dec 01 08:56:25 crc kubenswrapper[4873]: I1201 08:56:25.025391 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jvrlm" event={"ID":"439ab38e-6a5b-4aff-baef-0be711a2fad5","Type":"ContainerStarted","Data":"ed62d268071979319c6f5fd0301deaca9a27c66b540cb15525c4faa6d80a1bda"} Dec 01 08:56:25 crc kubenswrapper[4873]: I1201 08:56:25.047524 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tdc2x" podStartSLOduration=17.887890147 podStartE2EDuration="36.047491278s" podCreationTimestamp="2025-12-01 08:55:49 +0000 UTC" firstStartedPulling="2025-12-01 08:56:06.267723021 +0000 UTC m=+942.169831560" lastFinishedPulling="2025-12-01 08:56:24.427324152 +0000 UTC m=+960.329432691" observedRunningTime="2025-12-01 08:56:25.042473933 +0000 UTC m=+960.944582472" watchObservedRunningTime="2025-12-01 08:56:25.047491278 +0000 UTC m=+960.949599817" Dec 01 08:56:25 crc kubenswrapper[4873]: I1201 08:56:25.090435 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jvrlm" podStartSLOduration=12.068564092 podStartE2EDuration="30.090402706s" podCreationTimestamp="2025-12-01 08:55:55 +0000 UTC" firstStartedPulling="2025-12-01 08:56:06.453487382 +0000 UTC m=+942.355595921" lastFinishedPulling="2025-12-01 08:56:24.475325976 +0000 UTC m=+960.377434535" observedRunningTime="2025-12-01 08:56:25.08455968 +0000 UTC m=+960.986668219" watchObservedRunningTime="2025-12-01 08:56:25.090402706 +0000 UTC m=+960.992511245" Dec 01 08:56:25 crc kubenswrapper[4873]: I1201 08:56:25.111306 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-f6n4k" podStartSLOduration=13.687714237 podStartE2EDuration="32.111286285s" podCreationTimestamp="2025-12-01 08:55:53 +0000 UTC" firstStartedPulling="2025-12-01 08:56:06.166090352 +0000 UTC m=+942.068198891" lastFinishedPulling="2025-12-01 08:56:24.58966241 +0000 UTC m=+960.491770939" observedRunningTime="2025-12-01 08:56:25.109913961 +0000 UTC m=+961.012022520" watchObservedRunningTime="2025-12-01 08:56:25.111286285 +0000 UTC m=+961.013394824" Dec 01 08:56:26 crc kubenswrapper[4873]: I1201 08:56:26.017460 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jvrlm" Dec 01 08:56:26 crc kubenswrapper[4873]: I1201 08:56:26.017846 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jvrlm" Dec 01 08:56:27 crc kubenswrapper[4873]: I1201 08:56:27.070093 4873 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-jvrlm" podUID="439ab38e-6a5b-4aff-baef-0be711a2fad5" containerName="registry-server" probeResult="failure" output=< Dec 01 08:56:27 crc kubenswrapper[4873]: timeout: failed to connect service ":50051" within 1s Dec 01 08:56:27 crc kubenswrapper[4873]: > Dec 01 08:56:30 crc kubenswrapper[4873]: I1201 08:56:30.442429 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tdc2x" Dec 01 08:56:30 crc kubenswrapper[4873]: I1201 08:56:30.442491 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tdc2x" Dec 01 08:56:30 crc kubenswrapper[4873]: I1201 08:56:30.487008 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tdc2x" Dec 01 08:56:31 crc kubenswrapper[4873]: I1201 08:56:31.116236 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tdc2x" Dec 01 08:56:31 crc kubenswrapper[4873]: I1201 08:56:31.366265 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tdc2x"] Dec 01 08:56:33 crc kubenswrapper[4873]: I1201 08:56:33.079969 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-z2r2m" Dec 01 08:56:33 crc kubenswrapper[4873]: I1201 08:56:33.094172 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-tdc2x" podUID="77cdace7-aa56-4a6b-900b-b162ef0410a4" containerName="registry-server" containerID="cri-o://a70421f9712cd45f73b9a9ff27639f982bfbe63db77ce7c7f4477b42979c2d27" gracePeriod=2 Dec 01 08:56:33 crc kubenswrapper[4873]: I1201 08:56:33.141923 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-z2r2m" Dec 01 08:56:33 crc kubenswrapper[4873]: I1201 08:56:33.489791 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tdc2x" Dec 01 08:56:33 crc kubenswrapper[4873]: I1201 08:56:33.631371 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnlk5\" (UniqueName: \"kubernetes.io/projected/77cdace7-aa56-4a6b-900b-b162ef0410a4-kube-api-access-rnlk5\") pod \"77cdace7-aa56-4a6b-900b-b162ef0410a4\" (UID: \"77cdace7-aa56-4a6b-900b-b162ef0410a4\") " Dec 01 08:56:33 crc kubenswrapper[4873]: I1201 08:56:33.631505 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77cdace7-aa56-4a6b-900b-b162ef0410a4-utilities\") pod \"77cdace7-aa56-4a6b-900b-b162ef0410a4\" (UID: \"77cdace7-aa56-4a6b-900b-b162ef0410a4\") " Dec 01 08:56:33 crc kubenswrapper[4873]: I1201 08:56:33.631566 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77cdace7-aa56-4a6b-900b-b162ef0410a4-catalog-content\") pod \"77cdace7-aa56-4a6b-900b-b162ef0410a4\" (UID: \"77cdace7-aa56-4a6b-900b-b162ef0410a4\") " Dec 01 08:56:33 crc kubenswrapper[4873]: I1201 08:56:33.632440 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77cdace7-aa56-4a6b-900b-b162ef0410a4-utilities" (OuterVolumeSpecName: "utilities") pod "77cdace7-aa56-4a6b-900b-b162ef0410a4" (UID: "77cdace7-aa56-4a6b-900b-b162ef0410a4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:56:33 crc kubenswrapper[4873]: I1201 08:56:33.637991 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77cdace7-aa56-4a6b-900b-b162ef0410a4-kube-api-access-rnlk5" (OuterVolumeSpecName: "kube-api-access-rnlk5") pod "77cdace7-aa56-4a6b-900b-b162ef0410a4" (UID: "77cdace7-aa56-4a6b-900b-b162ef0410a4"). InnerVolumeSpecName "kube-api-access-rnlk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:56:33 crc kubenswrapper[4873]: I1201 08:56:33.681931 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77cdace7-aa56-4a6b-900b-b162ef0410a4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "77cdace7-aa56-4a6b-900b-b162ef0410a4" (UID: "77cdace7-aa56-4a6b-900b-b162ef0410a4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:56:33 crc kubenswrapper[4873]: I1201 08:56:33.700936 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-f6n4k" Dec 01 08:56:33 crc kubenswrapper[4873]: I1201 08:56:33.701039 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-f6n4k" Dec 01 08:56:33 crc kubenswrapper[4873]: I1201 08:56:33.732543 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnlk5\" (UniqueName: \"kubernetes.io/projected/77cdace7-aa56-4a6b-900b-b162ef0410a4-kube-api-access-rnlk5\") on node \"crc\" DevicePath \"\"" Dec 01 08:56:33 crc kubenswrapper[4873]: I1201 08:56:33.732597 4873 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77cdace7-aa56-4a6b-900b-b162ef0410a4-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:56:33 crc kubenswrapper[4873]: I1201 08:56:33.732609 4873 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77cdace7-aa56-4a6b-900b-b162ef0410a4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:56:33 crc kubenswrapper[4873]: I1201 08:56:33.758646 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-f6n4k" Dec 01 08:56:34 crc kubenswrapper[4873]: I1201 08:56:34.106068 4873 generic.go:334] "Generic (PLEG): container finished" podID="77cdace7-aa56-4a6b-900b-b162ef0410a4" containerID="a70421f9712cd45f73b9a9ff27639f982bfbe63db77ce7c7f4477b42979c2d27" exitCode=0 Dec 01 08:56:34 crc kubenswrapper[4873]: I1201 08:56:34.106216 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tdc2x" Dec 01 08:56:34 crc kubenswrapper[4873]: I1201 08:56:34.106184 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tdc2x" event={"ID":"77cdace7-aa56-4a6b-900b-b162ef0410a4","Type":"ContainerDied","Data":"a70421f9712cd45f73b9a9ff27639f982bfbe63db77ce7c7f4477b42979c2d27"} Dec 01 08:56:34 crc kubenswrapper[4873]: I1201 08:56:34.106291 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tdc2x" event={"ID":"77cdace7-aa56-4a6b-900b-b162ef0410a4","Type":"ContainerDied","Data":"0aa3a2f837ad59345f1a5a1b26d886cfb7d43963b971f621a2e0c5295777b1da"} Dec 01 08:56:34 crc kubenswrapper[4873]: I1201 08:56:34.106320 4873 scope.go:117] "RemoveContainer" containerID="a70421f9712cd45f73b9a9ff27639f982bfbe63db77ce7c7f4477b42979c2d27" Dec 01 08:56:34 crc kubenswrapper[4873]: I1201 08:56:34.126207 4873 scope.go:117] "RemoveContainer" containerID="f5daf524faa67d92cf347a0d236cacad6068db9373ed4987baa86d3d89d5f636" Dec 01 08:56:34 crc kubenswrapper[4873]: I1201 08:56:34.150214 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tdc2x"] Dec 01 08:56:34 crc kubenswrapper[4873]: I1201 08:56:34.159943 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-tdc2x"] Dec 01 08:56:34 crc kubenswrapper[4873]: I1201 08:56:34.166626 4873 scope.go:117] "RemoveContainer" containerID="13dd90155611dfdb26ad3895edb8ecab8b2a44a0d59464b7ffc4eda7ac39b901" Dec 01 08:56:34 crc kubenswrapper[4873]: I1201 08:56:34.167587 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-f6n4k" Dec 01 08:56:34 crc kubenswrapper[4873]: I1201 08:56:34.193352 4873 scope.go:117] "RemoveContainer" containerID="a70421f9712cd45f73b9a9ff27639f982bfbe63db77ce7c7f4477b42979c2d27" Dec 01 08:56:34 crc kubenswrapper[4873]: E1201 08:56:34.193980 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a70421f9712cd45f73b9a9ff27639f982bfbe63db77ce7c7f4477b42979c2d27\": container with ID starting with a70421f9712cd45f73b9a9ff27639f982bfbe63db77ce7c7f4477b42979c2d27 not found: ID does not exist" containerID="a70421f9712cd45f73b9a9ff27639f982bfbe63db77ce7c7f4477b42979c2d27" Dec 01 08:56:34 crc kubenswrapper[4873]: I1201 08:56:34.194060 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a70421f9712cd45f73b9a9ff27639f982bfbe63db77ce7c7f4477b42979c2d27"} err="failed to get container status \"a70421f9712cd45f73b9a9ff27639f982bfbe63db77ce7c7f4477b42979c2d27\": rpc error: code = NotFound desc = could not find container \"a70421f9712cd45f73b9a9ff27639f982bfbe63db77ce7c7f4477b42979c2d27\": container with ID starting with a70421f9712cd45f73b9a9ff27639f982bfbe63db77ce7c7f4477b42979c2d27 not found: ID does not exist" Dec 01 08:56:34 crc kubenswrapper[4873]: I1201 08:56:34.194135 4873 scope.go:117] "RemoveContainer" containerID="f5daf524faa67d92cf347a0d236cacad6068db9373ed4987baa86d3d89d5f636" Dec 01 08:56:34 crc kubenswrapper[4873]: E1201 08:56:34.194499 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5daf524faa67d92cf347a0d236cacad6068db9373ed4987baa86d3d89d5f636\": container with ID starting with f5daf524faa67d92cf347a0d236cacad6068db9373ed4987baa86d3d89d5f636 not found: ID does not exist" containerID="f5daf524faa67d92cf347a0d236cacad6068db9373ed4987baa86d3d89d5f636" Dec 01 08:56:34 crc kubenswrapper[4873]: I1201 08:56:34.194531 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5daf524faa67d92cf347a0d236cacad6068db9373ed4987baa86d3d89d5f636"} err="failed to get container status \"f5daf524faa67d92cf347a0d236cacad6068db9373ed4987baa86d3d89d5f636\": rpc error: code = NotFound desc = could not find container \"f5daf524faa67d92cf347a0d236cacad6068db9373ed4987baa86d3d89d5f636\": container with ID starting with f5daf524faa67d92cf347a0d236cacad6068db9373ed4987baa86d3d89d5f636 not found: ID does not exist" Dec 01 08:56:34 crc kubenswrapper[4873]: I1201 08:56:34.194562 4873 scope.go:117] "RemoveContainer" containerID="13dd90155611dfdb26ad3895edb8ecab8b2a44a0d59464b7ffc4eda7ac39b901" Dec 01 08:56:34 crc kubenswrapper[4873]: E1201 08:56:34.194813 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13dd90155611dfdb26ad3895edb8ecab8b2a44a0d59464b7ffc4eda7ac39b901\": container with ID starting with 13dd90155611dfdb26ad3895edb8ecab8b2a44a0d59464b7ffc4eda7ac39b901 not found: ID does not exist" containerID="13dd90155611dfdb26ad3895edb8ecab8b2a44a0d59464b7ffc4eda7ac39b901" Dec 01 08:56:34 crc kubenswrapper[4873]: I1201 08:56:34.194844 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13dd90155611dfdb26ad3895edb8ecab8b2a44a0d59464b7ffc4eda7ac39b901"} err="failed to get container status \"13dd90155611dfdb26ad3895edb8ecab8b2a44a0d59464b7ffc4eda7ac39b901\": rpc error: code = NotFound desc = could not find container \"13dd90155611dfdb26ad3895edb8ecab8b2a44a0d59464b7ffc4eda7ac39b901\": container with ID starting with 13dd90155611dfdb26ad3895edb8ecab8b2a44a0d59464b7ffc4eda7ac39b901 not found: ID does not exist" Dec 01 08:56:34 crc kubenswrapper[4873]: I1201 08:56:34.412596 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5rq8z" Dec 01 08:56:34 crc kubenswrapper[4873]: I1201 08:56:34.450097 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77cdace7-aa56-4a6b-900b-b162ef0410a4" path="/var/lib/kubelet/pods/77cdace7-aa56-4a6b-900b-b162ef0410a4/volumes" Dec 01 08:56:34 crc kubenswrapper[4873]: I1201 08:56:34.738540 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7wmqg" Dec 01 08:56:35 crc kubenswrapper[4873]: I1201 08:56:35.120611 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-tmdj7" event={"ID":"4dcc939e-b548-4fb6-814a-30e2aaa8a94a","Type":"ContainerStarted","Data":"75243be53c4db8331108f1406693625780f5f4a6b1888302b9993d517dc160c7"} Dec 01 08:56:35 crc kubenswrapper[4873]: I1201 08:56:35.121218 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-tmdj7" Dec 01 08:56:35 crc kubenswrapper[4873]: I1201 08:56:35.142299 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-57548d458d-tmdj7" podStartSLOduration=38.22409816 podStartE2EDuration="1m11.142271326s" podCreationTimestamp="2025-12-01 08:55:24 +0000 UTC" firstStartedPulling="2025-12-01 08:56:02.007228966 +0000 UTC m=+937.909337505" lastFinishedPulling="2025-12-01 08:56:34.925402132 +0000 UTC m=+970.827510671" observedRunningTime="2025-12-01 08:56:35.138045471 +0000 UTC m=+971.040154030" watchObservedRunningTime="2025-12-01 08:56:35.142271326 +0000 UTC m=+971.044379865" Dec 01 08:56:35 crc kubenswrapper[4873]: I1201 08:56:35.256395 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-hnzzf" Dec 01 08:56:35 crc kubenswrapper[4873]: I1201 08:56:35.341243 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-x9kc4" Dec 01 08:56:35 crc kubenswrapper[4873]: I1201 08:56:35.360327 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-z2r2m"] Dec 01 08:56:35 crc kubenswrapper[4873]: I1201 08:56:35.360652 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-z2r2m" podUID="f8f3966d-9400-44c1-a3e0-f4236541480e" containerName="registry-server" containerID="cri-o://42705b68e74d03f05c301a50433983cd22d82d47e640c48090dbef3efb037824" gracePeriod=2 Dec 01 08:56:35 crc kubenswrapper[4873]: I1201 08:56:35.368758 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-n79q9" Dec 01 08:56:35 crc kubenswrapper[4873]: I1201 08:56:35.611267 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-pfphw" Dec 01 08:56:35 crc kubenswrapper[4873]: I1201 08:56:35.766041 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z2r2m" Dec 01 08:56:35 crc kubenswrapper[4873]: I1201 08:56:35.867811 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4c6jw\" (UniqueName: \"kubernetes.io/projected/f8f3966d-9400-44c1-a3e0-f4236541480e-kube-api-access-4c6jw\") pod \"f8f3966d-9400-44c1-a3e0-f4236541480e\" (UID: \"f8f3966d-9400-44c1-a3e0-f4236541480e\") " Dec 01 08:56:35 crc kubenswrapper[4873]: I1201 08:56:35.867936 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8f3966d-9400-44c1-a3e0-f4236541480e-utilities\") pod \"f8f3966d-9400-44c1-a3e0-f4236541480e\" (UID: \"f8f3966d-9400-44c1-a3e0-f4236541480e\") " Dec 01 08:56:35 crc kubenswrapper[4873]: I1201 08:56:35.868008 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8f3966d-9400-44c1-a3e0-f4236541480e-catalog-content\") pod \"f8f3966d-9400-44c1-a3e0-f4236541480e\" (UID: \"f8f3966d-9400-44c1-a3e0-f4236541480e\") " Dec 01 08:56:35 crc kubenswrapper[4873]: I1201 08:56:35.870469 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8f3966d-9400-44c1-a3e0-f4236541480e-utilities" (OuterVolumeSpecName: "utilities") pod "f8f3966d-9400-44c1-a3e0-f4236541480e" (UID: "f8f3966d-9400-44c1-a3e0-f4236541480e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:56:35 crc kubenswrapper[4873]: I1201 08:56:35.874560 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8f3966d-9400-44c1-a3e0-f4236541480e-kube-api-access-4c6jw" (OuterVolumeSpecName: "kube-api-access-4c6jw") pod "f8f3966d-9400-44c1-a3e0-f4236541480e" (UID: "f8f3966d-9400-44c1-a3e0-f4236541480e"). InnerVolumeSpecName "kube-api-access-4c6jw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:56:35 crc kubenswrapper[4873]: I1201 08:56:35.937557 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-dx4jm" Dec 01 08:56:35 crc kubenswrapper[4873]: I1201 08:56:35.970117 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4c6jw\" (UniqueName: \"kubernetes.io/projected/f8f3966d-9400-44c1-a3e0-f4236541480e-kube-api-access-4c6jw\") on node \"crc\" DevicePath \"\"" Dec 01 08:56:35 crc kubenswrapper[4873]: I1201 08:56:35.970163 4873 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8f3966d-9400-44c1-a3e0-f4236541480e-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:56:35 crc kubenswrapper[4873]: I1201 08:56:35.982285 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8f3966d-9400-44c1-a3e0-f4236541480e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f8f3966d-9400-44c1-a3e0-f4236541480e" (UID: "f8f3966d-9400-44c1-a3e0-f4236541480e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:56:36 crc kubenswrapper[4873]: I1201 08:56:36.072280 4873 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8f3966d-9400-44c1-a3e0-f4236541480e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:56:36 crc kubenswrapper[4873]: I1201 08:56:36.087415 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jvrlm" Dec 01 08:56:36 crc kubenswrapper[4873]: I1201 08:56:36.133564 4873 generic.go:334] "Generic (PLEG): container finished" podID="f8f3966d-9400-44c1-a3e0-f4236541480e" containerID="42705b68e74d03f05c301a50433983cd22d82d47e640c48090dbef3efb037824" exitCode=0 Dec 01 08:56:36 crc kubenswrapper[4873]: I1201 08:56:36.133616 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z2r2m" event={"ID":"f8f3966d-9400-44c1-a3e0-f4236541480e","Type":"ContainerDied","Data":"42705b68e74d03f05c301a50433983cd22d82d47e640c48090dbef3efb037824"} Dec 01 08:56:36 crc kubenswrapper[4873]: I1201 08:56:36.133691 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z2r2m" event={"ID":"f8f3966d-9400-44c1-a3e0-f4236541480e","Type":"ContainerDied","Data":"dfca0c024b618172217edc647a9e04be8051e8782de00bba4e8281987d978456"} Dec 01 08:56:36 crc kubenswrapper[4873]: I1201 08:56:36.133696 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z2r2m" Dec 01 08:56:36 crc kubenswrapper[4873]: I1201 08:56:36.133719 4873 scope.go:117] "RemoveContainer" containerID="42705b68e74d03f05c301a50433983cd22d82d47e640c48090dbef3efb037824" Dec 01 08:56:36 crc kubenswrapper[4873]: I1201 08:56:36.138140 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jvrlm" Dec 01 08:56:36 crc kubenswrapper[4873]: I1201 08:56:36.153550 4873 scope.go:117] "RemoveContainer" containerID="978cabc5f4a7ff33f74ff41b164eb6ab71d1f0a87af985872823c682b3c75bcf" Dec 01 08:56:36 crc kubenswrapper[4873]: I1201 08:56:36.188160 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-z2r2m"] Dec 01 08:56:36 crc kubenswrapper[4873]: I1201 08:56:36.188900 4873 scope.go:117] "RemoveContainer" containerID="6a839b7303a1d01b67d8aca54815c2aa26443339aa86144c0a21b02b3ccac472" Dec 01 08:56:36 crc kubenswrapper[4873]: I1201 08:56:36.195392 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-z2r2m"] Dec 01 08:56:36 crc kubenswrapper[4873]: I1201 08:56:36.205619 4873 scope.go:117] "RemoveContainer" containerID="42705b68e74d03f05c301a50433983cd22d82d47e640c48090dbef3efb037824" Dec 01 08:56:36 crc kubenswrapper[4873]: E1201 08:56:36.206234 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42705b68e74d03f05c301a50433983cd22d82d47e640c48090dbef3efb037824\": container with ID starting with 42705b68e74d03f05c301a50433983cd22d82d47e640c48090dbef3efb037824 not found: ID does not exist" containerID="42705b68e74d03f05c301a50433983cd22d82d47e640c48090dbef3efb037824" Dec 01 08:56:36 crc kubenswrapper[4873]: I1201 08:56:36.206289 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42705b68e74d03f05c301a50433983cd22d82d47e640c48090dbef3efb037824"} err="failed to get container status \"42705b68e74d03f05c301a50433983cd22d82d47e640c48090dbef3efb037824\": rpc error: code = NotFound desc = could not find container \"42705b68e74d03f05c301a50433983cd22d82d47e640c48090dbef3efb037824\": container with ID starting with 42705b68e74d03f05c301a50433983cd22d82d47e640c48090dbef3efb037824 not found: ID does not exist" Dec 01 08:56:36 crc kubenswrapper[4873]: I1201 08:56:36.206390 4873 scope.go:117] "RemoveContainer" containerID="978cabc5f4a7ff33f74ff41b164eb6ab71d1f0a87af985872823c682b3c75bcf" Dec 01 08:56:36 crc kubenswrapper[4873]: E1201 08:56:36.206763 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"978cabc5f4a7ff33f74ff41b164eb6ab71d1f0a87af985872823c682b3c75bcf\": container with ID starting with 978cabc5f4a7ff33f74ff41b164eb6ab71d1f0a87af985872823c682b3c75bcf not found: ID does not exist" containerID="978cabc5f4a7ff33f74ff41b164eb6ab71d1f0a87af985872823c682b3c75bcf" Dec 01 08:56:36 crc kubenswrapper[4873]: I1201 08:56:36.206860 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"978cabc5f4a7ff33f74ff41b164eb6ab71d1f0a87af985872823c682b3c75bcf"} err="failed to get container status \"978cabc5f4a7ff33f74ff41b164eb6ab71d1f0a87af985872823c682b3c75bcf\": rpc error: code = NotFound desc = could not find container \"978cabc5f4a7ff33f74ff41b164eb6ab71d1f0a87af985872823c682b3c75bcf\": container with ID starting with 978cabc5f4a7ff33f74ff41b164eb6ab71d1f0a87af985872823c682b3c75bcf not found: ID does not exist" Dec 01 08:56:36 crc kubenswrapper[4873]: I1201 08:56:36.206893 4873 scope.go:117] "RemoveContainer" containerID="6a839b7303a1d01b67d8aca54815c2aa26443339aa86144c0a21b02b3ccac472" Dec 01 08:56:36 crc kubenswrapper[4873]: E1201 08:56:36.207427 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a839b7303a1d01b67d8aca54815c2aa26443339aa86144c0a21b02b3ccac472\": container with ID starting with 6a839b7303a1d01b67d8aca54815c2aa26443339aa86144c0a21b02b3ccac472 not found: ID does not exist" containerID="6a839b7303a1d01b67d8aca54815c2aa26443339aa86144c0a21b02b3ccac472" Dec 01 08:56:36 crc kubenswrapper[4873]: I1201 08:56:36.207458 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a839b7303a1d01b67d8aca54815c2aa26443339aa86144c0a21b02b3ccac472"} err="failed to get container status \"6a839b7303a1d01b67d8aca54815c2aa26443339aa86144c0a21b02b3ccac472\": rpc error: code = NotFound desc = could not find container \"6a839b7303a1d01b67d8aca54815c2aa26443339aa86144c0a21b02b3ccac472\": container with ID starting with 6a839b7303a1d01b67d8aca54815c2aa26443339aa86144c0a21b02b3ccac472 not found: ID does not exist" Dec 01 08:56:36 crc kubenswrapper[4873]: I1201 08:56:36.356631 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f6n4k"] Dec 01 08:56:36 crc kubenswrapper[4873]: I1201 08:56:36.356964 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-f6n4k" podUID="074c5f85-694d-47af-a33f-e1995ae73a5c" containerName="registry-server" containerID="cri-o://266158624fe4045657a423d2622eeda0d8688b0f02b87848d975907c902d88fb" gracePeriod=2 Dec 01 08:56:36 crc kubenswrapper[4873]: I1201 08:56:36.443329 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8f3966d-9400-44c1-a3e0-f4236541480e" path="/var/lib/kubelet/pods/f8f3966d-9400-44c1-a3e0-f4236541480e/volumes" Dec 01 08:56:36 crc kubenswrapper[4873]: I1201 08:56:36.792855 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f6n4k" Dec 01 08:56:36 crc kubenswrapper[4873]: I1201 08:56:36.886068 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/074c5f85-694d-47af-a33f-e1995ae73a5c-catalog-content\") pod \"074c5f85-694d-47af-a33f-e1995ae73a5c\" (UID: \"074c5f85-694d-47af-a33f-e1995ae73a5c\") " Dec 01 08:56:36 crc kubenswrapper[4873]: I1201 08:56:36.886259 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gkgsl\" (UniqueName: \"kubernetes.io/projected/074c5f85-694d-47af-a33f-e1995ae73a5c-kube-api-access-gkgsl\") pod \"074c5f85-694d-47af-a33f-e1995ae73a5c\" (UID: \"074c5f85-694d-47af-a33f-e1995ae73a5c\") " Dec 01 08:56:36 crc kubenswrapper[4873]: I1201 08:56:36.886359 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/074c5f85-694d-47af-a33f-e1995ae73a5c-utilities\") pod \"074c5f85-694d-47af-a33f-e1995ae73a5c\" (UID: \"074c5f85-694d-47af-a33f-e1995ae73a5c\") " Dec 01 08:56:36 crc kubenswrapper[4873]: I1201 08:56:36.887561 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/074c5f85-694d-47af-a33f-e1995ae73a5c-utilities" (OuterVolumeSpecName: "utilities") pod "074c5f85-694d-47af-a33f-e1995ae73a5c" (UID: "074c5f85-694d-47af-a33f-e1995ae73a5c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:56:36 crc kubenswrapper[4873]: I1201 08:56:36.891646 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/074c5f85-694d-47af-a33f-e1995ae73a5c-kube-api-access-gkgsl" (OuterVolumeSpecName: "kube-api-access-gkgsl") pod "074c5f85-694d-47af-a33f-e1995ae73a5c" (UID: "074c5f85-694d-47af-a33f-e1995ae73a5c"). InnerVolumeSpecName "kube-api-access-gkgsl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:56:36 crc kubenswrapper[4873]: I1201 08:56:36.905213 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/074c5f85-694d-47af-a33f-e1995ae73a5c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "074c5f85-694d-47af-a33f-e1995ae73a5c" (UID: "074c5f85-694d-47af-a33f-e1995ae73a5c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:56:36 crc kubenswrapper[4873]: I1201 08:56:36.988846 4873 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/074c5f85-694d-47af-a33f-e1995ae73a5c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:56:36 crc kubenswrapper[4873]: I1201 08:56:36.988891 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gkgsl\" (UniqueName: \"kubernetes.io/projected/074c5f85-694d-47af-a33f-e1995ae73a5c-kube-api-access-gkgsl\") on node \"crc\" DevicePath \"\"" Dec 01 08:56:36 crc kubenswrapper[4873]: I1201 08:56:36.988905 4873 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/074c5f85-694d-47af-a33f-e1995ae73a5c-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:56:37 crc kubenswrapper[4873]: I1201 08:56:37.150128 4873 generic.go:334] "Generic (PLEG): container finished" podID="074c5f85-694d-47af-a33f-e1995ae73a5c" containerID="266158624fe4045657a423d2622eeda0d8688b0f02b87848d975907c902d88fb" exitCode=0 Dec 01 08:56:37 crc kubenswrapper[4873]: I1201 08:56:37.150258 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f6n4k" Dec 01 08:56:37 crc kubenswrapper[4873]: I1201 08:56:37.150226 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f6n4k" event={"ID":"074c5f85-694d-47af-a33f-e1995ae73a5c","Type":"ContainerDied","Data":"266158624fe4045657a423d2622eeda0d8688b0f02b87848d975907c902d88fb"} Dec 01 08:56:37 crc kubenswrapper[4873]: I1201 08:56:37.150518 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f6n4k" event={"ID":"074c5f85-694d-47af-a33f-e1995ae73a5c","Type":"ContainerDied","Data":"1fc588e35318ef8b19cd182fddcefe703211662ebe652497c89196c3191cfe7e"} Dec 01 08:56:37 crc kubenswrapper[4873]: I1201 08:56:37.150555 4873 scope.go:117] "RemoveContainer" containerID="266158624fe4045657a423d2622eeda0d8688b0f02b87848d975907c902d88fb" Dec 01 08:56:37 crc kubenswrapper[4873]: I1201 08:56:37.169112 4873 scope.go:117] "RemoveContainer" containerID="d6b1c3d0d1b343c324e7107882e34aeced110702cd6413eb2a36c9606e32d77b" Dec 01 08:56:37 crc kubenswrapper[4873]: I1201 08:56:37.188941 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f6n4k"] Dec 01 08:56:37 crc kubenswrapper[4873]: I1201 08:56:37.196944 4873 scope.go:117] "RemoveContainer" containerID="ae86de4f1adff7a38a04222ea6228b7ecf71a389fd4a05dfe9c5150cc951b5b5" Dec 01 08:56:37 crc kubenswrapper[4873]: I1201 08:56:37.197375 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-f6n4k"] Dec 01 08:56:37 crc kubenswrapper[4873]: I1201 08:56:37.213644 4873 scope.go:117] "RemoveContainer" containerID="266158624fe4045657a423d2622eeda0d8688b0f02b87848d975907c902d88fb" Dec 01 08:56:37 crc kubenswrapper[4873]: E1201 08:56:37.213969 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"266158624fe4045657a423d2622eeda0d8688b0f02b87848d975907c902d88fb\": container with ID starting with 266158624fe4045657a423d2622eeda0d8688b0f02b87848d975907c902d88fb not found: ID does not exist" containerID="266158624fe4045657a423d2622eeda0d8688b0f02b87848d975907c902d88fb" Dec 01 08:56:37 crc kubenswrapper[4873]: I1201 08:56:37.214005 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"266158624fe4045657a423d2622eeda0d8688b0f02b87848d975907c902d88fb"} err="failed to get container status \"266158624fe4045657a423d2622eeda0d8688b0f02b87848d975907c902d88fb\": rpc error: code = NotFound desc = could not find container \"266158624fe4045657a423d2622eeda0d8688b0f02b87848d975907c902d88fb\": container with ID starting with 266158624fe4045657a423d2622eeda0d8688b0f02b87848d975907c902d88fb not found: ID does not exist" Dec 01 08:56:37 crc kubenswrapper[4873]: I1201 08:56:37.214047 4873 scope.go:117] "RemoveContainer" containerID="d6b1c3d0d1b343c324e7107882e34aeced110702cd6413eb2a36c9606e32d77b" Dec 01 08:56:37 crc kubenswrapper[4873]: E1201 08:56:37.214356 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6b1c3d0d1b343c324e7107882e34aeced110702cd6413eb2a36c9606e32d77b\": container with ID starting with d6b1c3d0d1b343c324e7107882e34aeced110702cd6413eb2a36c9606e32d77b not found: ID does not exist" containerID="d6b1c3d0d1b343c324e7107882e34aeced110702cd6413eb2a36c9606e32d77b" Dec 01 08:56:37 crc kubenswrapper[4873]: I1201 08:56:37.214379 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6b1c3d0d1b343c324e7107882e34aeced110702cd6413eb2a36c9606e32d77b"} err="failed to get container status \"d6b1c3d0d1b343c324e7107882e34aeced110702cd6413eb2a36c9606e32d77b\": rpc error: code = NotFound desc = could not find container \"d6b1c3d0d1b343c324e7107882e34aeced110702cd6413eb2a36c9606e32d77b\": container with ID starting with d6b1c3d0d1b343c324e7107882e34aeced110702cd6413eb2a36c9606e32d77b not found: ID does not exist" Dec 01 08:56:37 crc kubenswrapper[4873]: I1201 08:56:37.214399 4873 scope.go:117] "RemoveContainer" containerID="ae86de4f1adff7a38a04222ea6228b7ecf71a389fd4a05dfe9c5150cc951b5b5" Dec 01 08:56:37 crc kubenswrapper[4873]: E1201 08:56:37.214617 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae86de4f1adff7a38a04222ea6228b7ecf71a389fd4a05dfe9c5150cc951b5b5\": container with ID starting with ae86de4f1adff7a38a04222ea6228b7ecf71a389fd4a05dfe9c5150cc951b5b5 not found: ID does not exist" containerID="ae86de4f1adff7a38a04222ea6228b7ecf71a389fd4a05dfe9c5150cc951b5b5" Dec 01 08:56:37 crc kubenswrapper[4873]: I1201 08:56:37.214644 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae86de4f1adff7a38a04222ea6228b7ecf71a389fd4a05dfe9c5150cc951b5b5"} err="failed to get container status \"ae86de4f1adff7a38a04222ea6228b7ecf71a389fd4a05dfe9c5150cc951b5b5\": rpc error: code = NotFound desc = could not find container \"ae86de4f1adff7a38a04222ea6228b7ecf71a389fd4a05dfe9c5150cc951b5b5\": container with ID starting with ae86de4f1adff7a38a04222ea6228b7ecf71a389fd4a05dfe9c5150cc951b5b5 not found: ID does not exist" Dec 01 08:56:37 crc kubenswrapper[4873]: E1201 08:56:37.432160 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\"" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6dxcw" podUID="f7a7bcca-9403-4e2d-bc28-53d4eb5ae252" Dec 01 08:56:38 crc kubenswrapper[4873]: I1201 08:56:38.167761 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vf8qt" event={"ID":"5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5","Type":"ContainerStarted","Data":"78c3c2ed689eb1e481bfdf541672799637adc25b98b0f70e079083ced487fe32"} Dec 01 08:56:38 crc kubenswrapper[4873]: I1201 08:56:38.168043 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vf8qt" Dec 01 08:56:38 crc kubenswrapper[4873]: I1201 08:56:38.197661 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vf8qt" podStartSLOduration=40.545332136 podStartE2EDuration="1m14.197632785s" podCreationTimestamp="2025-12-01 08:55:24 +0000 UTC" firstStartedPulling="2025-12-01 08:56:03.343350572 +0000 UTC m=+939.245459111" lastFinishedPulling="2025-12-01 08:56:36.995651221 +0000 UTC m=+972.897759760" observedRunningTime="2025-12-01 08:56:38.193029261 +0000 UTC m=+974.095137810" watchObservedRunningTime="2025-12-01 08:56:38.197632785 +0000 UTC m=+974.099741334" Dec 01 08:56:38 crc kubenswrapper[4873]: I1201 08:56:38.448980 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="074c5f85-694d-47af-a33f-e1995ae73a5c" path="/var/lib/kubelet/pods/074c5f85-694d-47af-a33f-e1995ae73a5c/volumes" Dec 01 08:56:39 crc kubenswrapper[4873]: I1201 08:56:39.757353 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jvrlm"] Dec 01 08:56:39 crc kubenswrapper[4873]: I1201 08:56:39.757986 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jvrlm" podUID="439ab38e-6a5b-4aff-baef-0be711a2fad5" containerName="registry-server" containerID="cri-o://ed62d268071979319c6f5fd0301deaca9a27c66b540cb15525c4faa6d80a1bda" gracePeriod=2 Dec 01 08:56:40 crc kubenswrapper[4873]: I1201 08:56:40.191811 4873 generic.go:334] "Generic (PLEG): container finished" podID="439ab38e-6a5b-4aff-baef-0be711a2fad5" containerID="ed62d268071979319c6f5fd0301deaca9a27c66b540cb15525c4faa6d80a1bda" exitCode=0 Dec 01 08:56:40 crc kubenswrapper[4873]: I1201 08:56:40.191896 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jvrlm" event={"ID":"439ab38e-6a5b-4aff-baef-0be711a2fad5","Type":"ContainerDied","Data":"ed62d268071979319c6f5fd0301deaca9a27c66b540cb15525c4faa6d80a1bda"} Dec 01 08:56:40 crc kubenswrapper[4873]: I1201 08:56:40.233934 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jvrlm" Dec 01 08:56:40 crc kubenswrapper[4873]: I1201 08:56:40.346511 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/439ab38e-6a5b-4aff-baef-0be711a2fad5-catalog-content\") pod \"439ab38e-6a5b-4aff-baef-0be711a2fad5\" (UID: \"439ab38e-6a5b-4aff-baef-0be711a2fad5\") " Dec 01 08:56:40 crc kubenswrapper[4873]: I1201 08:56:40.346678 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/439ab38e-6a5b-4aff-baef-0be711a2fad5-utilities\") pod \"439ab38e-6a5b-4aff-baef-0be711a2fad5\" (UID: \"439ab38e-6a5b-4aff-baef-0be711a2fad5\") " Dec 01 08:56:40 crc kubenswrapper[4873]: I1201 08:56:40.346747 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdrp8\" (UniqueName: \"kubernetes.io/projected/439ab38e-6a5b-4aff-baef-0be711a2fad5-kube-api-access-kdrp8\") pod \"439ab38e-6a5b-4aff-baef-0be711a2fad5\" (UID: \"439ab38e-6a5b-4aff-baef-0be711a2fad5\") " Dec 01 08:56:40 crc kubenswrapper[4873]: I1201 08:56:40.347772 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/439ab38e-6a5b-4aff-baef-0be711a2fad5-utilities" (OuterVolumeSpecName: "utilities") pod "439ab38e-6a5b-4aff-baef-0be711a2fad5" (UID: "439ab38e-6a5b-4aff-baef-0be711a2fad5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:56:40 crc kubenswrapper[4873]: I1201 08:56:40.352317 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/439ab38e-6a5b-4aff-baef-0be711a2fad5-kube-api-access-kdrp8" (OuterVolumeSpecName: "kube-api-access-kdrp8") pod "439ab38e-6a5b-4aff-baef-0be711a2fad5" (UID: "439ab38e-6a5b-4aff-baef-0be711a2fad5"). InnerVolumeSpecName "kube-api-access-kdrp8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:56:40 crc kubenswrapper[4873]: I1201 08:56:40.400198 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/439ab38e-6a5b-4aff-baef-0be711a2fad5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "439ab38e-6a5b-4aff-baef-0be711a2fad5" (UID: "439ab38e-6a5b-4aff-baef-0be711a2fad5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:56:40 crc kubenswrapper[4873]: I1201 08:56:40.448834 4873 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/439ab38e-6a5b-4aff-baef-0be711a2fad5-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:56:40 crc kubenswrapper[4873]: I1201 08:56:40.448879 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kdrp8\" (UniqueName: \"kubernetes.io/projected/439ab38e-6a5b-4aff-baef-0be711a2fad5-kube-api-access-kdrp8\") on node \"crc\" DevicePath \"\"" Dec 01 08:56:40 crc kubenswrapper[4873]: I1201 08:56:40.448902 4873 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/439ab38e-6a5b-4aff-baef-0be711a2fad5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:56:40 crc kubenswrapper[4873]: I1201 08:56:40.629631 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-tmdj7" Dec 01 08:56:41 crc kubenswrapper[4873]: I1201 08:56:41.204262 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jvrlm" event={"ID":"439ab38e-6a5b-4aff-baef-0be711a2fad5","Type":"ContainerDied","Data":"c2d37e795ab7413bbc0160cf511a0bf8ce9e930f4abfa1bcd5ac8654cbccb517"} Dec 01 08:56:41 crc kubenswrapper[4873]: I1201 08:56:41.204348 4873 scope.go:117] "RemoveContainer" containerID="ed62d268071979319c6f5fd0301deaca9a27c66b540cb15525c4faa6d80a1bda" Dec 01 08:56:41 crc kubenswrapper[4873]: I1201 08:56:41.204375 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jvrlm" Dec 01 08:56:41 crc kubenswrapper[4873]: I1201 08:56:41.237766 4873 scope.go:117] "RemoveContainer" containerID="ad585e8259e362201e754008e702c73364e7a1a57452d275026cb18e63c71963" Dec 01 08:56:41 crc kubenswrapper[4873]: I1201 08:56:41.252770 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jvrlm"] Dec 01 08:56:41 crc kubenswrapper[4873]: I1201 08:56:41.263843 4873 scope.go:117] "RemoveContainer" containerID="a9c04480d5d9be7bc887ae98d780d9b51325cf96f9a0c3806a4a1d9156bbaafa" Dec 01 08:56:41 crc kubenswrapper[4873]: I1201 08:56:41.270870 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jvrlm"] Dec 01 08:56:42 crc kubenswrapper[4873]: I1201 08:56:42.447222 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="439ab38e-6a5b-4aff-baef-0be711a2fad5" path="/var/lib/kubelet/pods/439ab38e-6a5b-4aff-baef-0be711a2fad5/volumes" Dec 01 08:56:51 crc kubenswrapper[4873]: I1201 08:56:51.057722 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vf8qt" Dec 01 08:56:52 crc kubenswrapper[4873]: I1201 08:56:52.323241 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6dxcw" event={"ID":"f7a7bcca-9403-4e2d-bc28-53d4eb5ae252","Type":"ContainerStarted","Data":"7e1790666a9f5a34f04715d2060b93da0289f81f173a48e3ac4f697fcfed6d50"} Dec 01 08:56:52 crc kubenswrapper[4873]: I1201 08:56:52.324078 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6dxcw" Dec 01 08:56:52 crc kubenswrapper[4873]: I1201 08:56:52.343629 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6dxcw" podStartSLOduration=5.088399787 podStartE2EDuration="1m28.343596626s" podCreationTimestamp="2025-12-01 08:55:24 +0000 UTC" firstStartedPulling="2025-12-01 08:55:28.71487019 +0000 UTC m=+904.616978729" lastFinishedPulling="2025-12-01 08:56:51.970067039 +0000 UTC m=+987.872175568" observedRunningTime="2025-12-01 08:56:52.338452788 +0000 UTC m=+988.240561357" watchObservedRunningTime="2025-12-01 08:56:52.343596626 +0000 UTC m=+988.245705185" Dec 01 08:57:05 crc kubenswrapper[4873]: I1201 08:57:05.815772 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6dxcw" Dec 01 08:57:19 crc kubenswrapper[4873]: I1201 08:57:19.692742 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-jxpr9"] Dec 01 08:57:19 crc kubenswrapper[4873]: E1201 08:57:19.693877 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77cdace7-aa56-4a6b-900b-b162ef0410a4" containerName="extract-utilities" Dec 01 08:57:19 crc kubenswrapper[4873]: I1201 08:57:19.693896 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="77cdace7-aa56-4a6b-900b-b162ef0410a4" containerName="extract-utilities" Dec 01 08:57:19 crc kubenswrapper[4873]: E1201 08:57:19.693927 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="439ab38e-6a5b-4aff-baef-0be711a2fad5" containerName="extract-content" Dec 01 08:57:19 crc kubenswrapper[4873]: I1201 08:57:19.693934 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="439ab38e-6a5b-4aff-baef-0be711a2fad5" containerName="extract-content" Dec 01 08:57:19 crc kubenswrapper[4873]: E1201 08:57:19.693943 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="074c5f85-694d-47af-a33f-e1995ae73a5c" containerName="extract-content" Dec 01 08:57:19 crc kubenswrapper[4873]: I1201 08:57:19.693950 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="074c5f85-694d-47af-a33f-e1995ae73a5c" containerName="extract-content" Dec 01 08:57:19 crc kubenswrapper[4873]: E1201 08:57:19.693965 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77cdace7-aa56-4a6b-900b-b162ef0410a4" containerName="extract-content" Dec 01 08:57:19 crc kubenswrapper[4873]: I1201 08:57:19.693971 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="77cdace7-aa56-4a6b-900b-b162ef0410a4" containerName="extract-content" Dec 01 08:57:19 crc kubenswrapper[4873]: E1201 08:57:19.693989 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="074c5f85-694d-47af-a33f-e1995ae73a5c" containerName="extract-utilities" Dec 01 08:57:19 crc kubenswrapper[4873]: I1201 08:57:19.693995 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="074c5f85-694d-47af-a33f-e1995ae73a5c" containerName="extract-utilities" Dec 01 08:57:19 crc kubenswrapper[4873]: E1201 08:57:19.694005 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8f3966d-9400-44c1-a3e0-f4236541480e" containerName="extract-content" Dec 01 08:57:19 crc kubenswrapper[4873]: I1201 08:57:19.694027 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8f3966d-9400-44c1-a3e0-f4236541480e" containerName="extract-content" Dec 01 08:57:19 crc kubenswrapper[4873]: E1201 08:57:19.694040 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8f3966d-9400-44c1-a3e0-f4236541480e" containerName="registry-server" Dec 01 08:57:19 crc kubenswrapper[4873]: I1201 08:57:19.694046 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8f3966d-9400-44c1-a3e0-f4236541480e" containerName="registry-server" Dec 01 08:57:19 crc kubenswrapper[4873]: E1201 08:57:19.694057 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77cdace7-aa56-4a6b-900b-b162ef0410a4" containerName="registry-server" Dec 01 08:57:19 crc kubenswrapper[4873]: I1201 08:57:19.694065 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="77cdace7-aa56-4a6b-900b-b162ef0410a4" containerName="registry-server" Dec 01 08:57:19 crc kubenswrapper[4873]: E1201 08:57:19.694081 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="439ab38e-6a5b-4aff-baef-0be711a2fad5" containerName="extract-utilities" Dec 01 08:57:19 crc kubenswrapper[4873]: I1201 08:57:19.694087 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="439ab38e-6a5b-4aff-baef-0be711a2fad5" containerName="extract-utilities" Dec 01 08:57:19 crc kubenswrapper[4873]: E1201 08:57:19.694096 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8f3966d-9400-44c1-a3e0-f4236541480e" containerName="extract-utilities" Dec 01 08:57:19 crc kubenswrapper[4873]: I1201 08:57:19.694101 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8f3966d-9400-44c1-a3e0-f4236541480e" containerName="extract-utilities" Dec 01 08:57:19 crc kubenswrapper[4873]: E1201 08:57:19.694112 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="074c5f85-694d-47af-a33f-e1995ae73a5c" containerName="registry-server" Dec 01 08:57:19 crc kubenswrapper[4873]: I1201 08:57:19.694118 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="074c5f85-694d-47af-a33f-e1995ae73a5c" containerName="registry-server" Dec 01 08:57:19 crc kubenswrapper[4873]: E1201 08:57:19.694129 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="439ab38e-6a5b-4aff-baef-0be711a2fad5" containerName="registry-server" Dec 01 08:57:19 crc kubenswrapper[4873]: I1201 08:57:19.694135 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="439ab38e-6a5b-4aff-baef-0be711a2fad5" containerName="registry-server" Dec 01 08:57:19 crc kubenswrapper[4873]: I1201 08:57:19.694291 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="439ab38e-6a5b-4aff-baef-0be711a2fad5" containerName="registry-server" Dec 01 08:57:19 crc kubenswrapper[4873]: I1201 08:57:19.694304 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="074c5f85-694d-47af-a33f-e1995ae73a5c" containerName="registry-server" Dec 01 08:57:19 crc kubenswrapper[4873]: I1201 08:57:19.694316 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="77cdace7-aa56-4a6b-900b-b162ef0410a4" containerName="registry-server" Dec 01 08:57:19 crc kubenswrapper[4873]: I1201 08:57:19.694330 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8f3966d-9400-44c1-a3e0-f4236541480e" containerName="registry-server" Dec 01 08:57:19 crc kubenswrapper[4873]: I1201 08:57:19.696515 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-jxpr9" Dec 01 08:57:19 crc kubenswrapper[4873]: I1201 08:57:19.705172 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-jxpr9"] Dec 01 08:57:19 crc kubenswrapper[4873]: I1201 08:57:19.707189 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-wvbn2" Dec 01 08:57:19 crc kubenswrapper[4873]: I1201 08:57:19.708004 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 01 08:57:19 crc kubenswrapper[4873]: I1201 08:57:19.708158 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 01 08:57:19 crc kubenswrapper[4873]: I1201 08:57:19.708219 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 01 08:57:19 crc kubenswrapper[4873]: I1201 08:57:19.764552 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-hssdh"] Dec 01 08:57:19 crc kubenswrapper[4873]: I1201 08:57:19.766110 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-hssdh" Dec 01 08:57:19 crc kubenswrapper[4873]: I1201 08:57:19.769640 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 01 08:57:19 crc kubenswrapper[4873]: I1201 08:57:19.779063 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-hssdh"] Dec 01 08:57:19 crc kubenswrapper[4873]: I1201 08:57:19.867224 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a6dc36f-94eb-45a1-b0cd-4f351cd01b1e-config\") pod \"dnsmasq-dns-78dd6ddcc-hssdh\" (UID: \"2a6dc36f-94eb-45a1-b0cd-4f351cd01b1e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-hssdh" Dec 01 08:57:19 crc kubenswrapper[4873]: I1201 08:57:19.867320 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-449gp\" (UniqueName: \"kubernetes.io/projected/4e867ef7-6308-48b5-b070-fa0ba4513e7f-kube-api-access-449gp\") pod \"dnsmasq-dns-675f4bcbfc-jxpr9\" (UID: \"4e867ef7-6308-48b5-b070-fa0ba4513e7f\") " pod="openstack/dnsmasq-dns-675f4bcbfc-jxpr9" Dec 01 08:57:19 crc kubenswrapper[4873]: I1201 08:57:19.867360 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e867ef7-6308-48b5-b070-fa0ba4513e7f-config\") pod \"dnsmasq-dns-675f4bcbfc-jxpr9\" (UID: \"4e867ef7-6308-48b5-b070-fa0ba4513e7f\") " pod="openstack/dnsmasq-dns-675f4bcbfc-jxpr9" Dec 01 08:57:19 crc kubenswrapper[4873]: I1201 08:57:19.867462 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2a6dc36f-94eb-45a1-b0cd-4f351cd01b1e-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-hssdh\" (UID: \"2a6dc36f-94eb-45a1-b0cd-4f351cd01b1e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-hssdh" Dec 01 08:57:19 crc kubenswrapper[4873]: I1201 08:57:19.867487 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rl9bf\" (UniqueName: \"kubernetes.io/projected/2a6dc36f-94eb-45a1-b0cd-4f351cd01b1e-kube-api-access-rl9bf\") pod \"dnsmasq-dns-78dd6ddcc-hssdh\" (UID: \"2a6dc36f-94eb-45a1-b0cd-4f351cd01b1e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-hssdh" Dec 01 08:57:19 crc kubenswrapper[4873]: I1201 08:57:19.968996 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-449gp\" (UniqueName: \"kubernetes.io/projected/4e867ef7-6308-48b5-b070-fa0ba4513e7f-kube-api-access-449gp\") pod \"dnsmasq-dns-675f4bcbfc-jxpr9\" (UID: \"4e867ef7-6308-48b5-b070-fa0ba4513e7f\") " pod="openstack/dnsmasq-dns-675f4bcbfc-jxpr9" Dec 01 08:57:19 crc kubenswrapper[4873]: I1201 08:57:19.969418 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e867ef7-6308-48b5-b070-fa0ba4513e7f-config\") pod \"dnsmasq-dns-675f4bcbfc-jxpr9\" (UID: \"4e867ef7-6308-48b5-b070-fa0ba4513e7f\") " pod="openstack/dnsmasq-dns-675f4bcbfc-jxpr9" Dec 01 08:57:19 crc kubenswrapper[4873]: I1201 08:57:19.969488 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2a6dc36f-94eb-45a1-b0cd-4f351cd01b1e-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-hssdh\" (UID: \"2a6dc36f-94eb-45a1-b0cd-4f351cd01b1e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-hssdh" Dec 01 08:57:19 crc kubenswrapper[4873]: I1201 08:57:19.969523 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rl9bf\" (UniqueName: \"kubernetes.io/projected/2a6dc36f-94eb-45a1-b0cd-4f351cd01b1e-kube-api-access-rl9bf\") pod \"dnsmasq-dns-78dd6ddcc-hssdh\" (UID: \"2a6dc36f-94eb-45a1-b0cd-4f351cd01b1e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-hssdh" Dec 01 08:57:19 crc kubenswrapper[4873]: I1201 08:57:19.969545 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a6dc36f-94eb-45a1-b0cd-4f351cd01b1e-config\") pod \"dnsmasq-dns-78dd6ddcc-hssdh\" (UID: \"2a6dc36f-94eb-45a1-b0cd-4f351cd01b1e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-hssdh" Dec 01 08:57:19 crc kubenswrapper[4873]: I1201 08:57:19.970585 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a6dc36f-94eb-45a1-b0cd-4f351cd01b1e-config\") pod \"dnsmasq-dns-78dd6ddcc-hssdh\" (UID: \"2a6dc36f-94eb-45a1-b0cd-4f351cd01b1e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-hssdh" Dec 01 08:57:19 crc kubenswrapper[4873]: I1201 08:57:19.971597 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e867ef7-6308-48b5-b070-fa0ba4513e7f-config\") pod \"dnsmasq-dns-675f4bcbfc-jxpr9\" (UID: \"4e867ef7-6308-48b5-b070-fa0ba4513e7f\") " pod="openstack/dnsmasq-dns-675f4bcbfc-jxpr9" Dec 01 08:57:19 crc kubenswrapper[4873]: I1201 08:57:19.972204 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2a6dc36f-94eb-45a1-b0cd-4f351cd01b1e-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-hssdh\" (UID: \"2a6dc36f-94eb-45a1-b0cd-4f351cd01b1e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-hssdh" Dec 01 08:57:19 crc kubenswrapper[4873]: I1201 08:57:19.999358 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rl9bf\" (UniqueName: \"kubernetes.io/projected/2a6dc36f-94eb-45a1-b0cd-4f351cd01b1e-kube-api-access-rl9bf\") pod \"dnsmasq-dns-78dd6ddcc-hssdh\" (UID: \"2a6dc36f-94eb-45a1-b0cd-4f351cd01b1e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-hssdh" Dec 01 08:57:20 crc kubenswrapper[4873]: I1201 08:57:20.000041 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-449gp\" (UniqueName: \"kubernetes.io/projected/4e867ef7-6308-48b5-b070-fa0ba4513e7f-kube-api-access-449gp\") pod \"dnsmasq-dns-675f4bcbfc-jxpr9\" (UID: \"4e867ef7-6308-48b5-b070-fa0ba4513e7f\") " pod="openstack/dnsmasq-dns-675f4bcbfc-jxpr9" Dec 01 08:57:20 crc kubenswrapper[4873]: I1201 08:57:20.035777 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-jxpr9" Dec 01 08:57:20 crc kubenswrapper[4873]: I1201 08:57:20.086703 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-hssdh" Dec 01 08:57:20 crc kubenswrapper[4873]: I1201 08:57:20.402265 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-hssdh"] Dec 01 08:57:20 crc kubenswrapper[4873]: W1201 08:57:20.506439 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4e867ef7_6308_48b5_b070_fa0ba4513e7f.slice/crio-745447136c68f7001636207498c3efaecaaee554488c7b7294ff861975dcc754 WatchSource:0}: Error finding container 745447136c68f7001636207498c3efaecaaee554488c7b7294ff861975dcc754: Status 404 returned error can't find the container with id 745447136c68f7001636207498c3efaecaaee554488c7b7294ff861975dcc754 Dec 01 08:57:20 crc kubenswrapper[4873]: I1201 08:57:20.507033 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-jxpr9"] Dec 01 08:57:20 crc kubenswrapper[4873]: I1201 08:57:20.612159 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-hssdh" event={"ID":"2a6dc36f-94eb-45a1-b0cd-4f351cd01b1e","Type":"ContainerStarted","Data":"d7b3e03ad2a06029fad081c99cdcddee4d95212b11212a4044e95a42f693ade3"} Dec 01 08:57:20 crc kubenswrapper[4873]: I1201 08:57:20.613492 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-jxpr9" event={"ID":"4e867ef7-6308-48b5-b070-fa0ba4513e7f","Type":"ContainerStarted","Data":"745447136c68f7001636207498c3efaecaaee554488c7b7294ff861975dcc754"} Dec 01 08:57:22 crc kubenswrapper[4873]: I1201 08:57:22.714385 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-jxpr9"] Dec 01 08:57:22 crc kubenswrapper[4873]: I1201 08:57:22.748522 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-gpkrq"] Dec 01 08:57:22 crc kubenswrapper[4873]: I1201 08:57:22.749901 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-gpkrq" Dec 01 08:57:22 crc kubenswrapper[4873]: I1201 08:57:22.776826 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-gpkrq"] Dec 01 08:57:22 crc kubenswrapper[4873]: I1201 08:57:22.821639 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/34cbfd49-8920-4490-9f8a-7a6f2542d1c6-dns-svc\") pod \"dnsmasq-dns-666b6646f7-gpkrq\" (UID: \"34cbfd49-8920-4490-9f8a-7a6f2542d1c6\") " pod="openstack/dnsmasq-dns-666b6646f7-gpkrq" Dec 01 08:57:22 crc kubenswrapper[4873]: I1201 08:57:22.821743 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tzzd\" (UniqueName: \"kubernetes.io/projected/34cbfd49-8920-4490-9f8a-7a6f2542d1c6-kube-api-access-5tzzd\") pod \"dnsmasq-dns-666b6646f7-gpkrq\" (UID: \"34cbfd49-8920-4490-9f8a-7a6f2542d1c6\") " pod="openstack/dnsmasq-dns-666b6646f7-gpkrq" Dec 01 08:57:22 crc kubenswrapper[4873]: I1201 08:57:22.821818 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34cbfd49-8920-4490-9f8a-7a6f2542d1c6-config\") pod \"dnsmasq-dns-666b6646f7-gpkrq\" (UID: \"34cbfd49-8920-4490-9f8a-7a6f2542d1c6\") " pod="openstack/dnsmasq-dns-666b6646f7-gpkrq" Dec 01 08:57:22 crc kubenswrapper[4873]: I1201 08:57:22.924287 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/34cbfd49-8920-4490-9f8a-7a6f2542d1c6-dns-svc\") pod \"dnsmasq-dns-666b6646f7-gpkrq\" (UID: \"34cbfd49-8920-4490-9f8a-7a6f2542d1c6\") " pod="openstack/dnsmasq-dns-666b6646f7-gpkrq" Dec 01 08:57:22 crc kubenswrapper[4873]: I1201 08:57:22.924412 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tzzd\" (UniqueName: \"kubernetes.io/projected/34cbfd49-8920-4490-9f8a-7a6f2542d1c6-kube-api-access-5tzzd\") pod \"dnsmasq-dns-666b6646f7-gpkrq\" (UID: \"34cbfd49-8920-4490-9f8a-7a6f2542d1c6\") " pod="openstack/dnsmasq-dns-666b6646f7-gpkrq" Dec 01 08:57:22 crc kubenswrapper[4873]: I1201 08:57:22.924504 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34cbfd49-8920-4490-9f8a-7a6f2542d1c6-config\") pod \"dnsmasq-dns-666b6646f7-gpkrq\" (UID: \"34cbfd49-8920-4490-9f8a-7a6f2542d1c6\") " pod="openstack/dnsmasq-dns-666b6646f7-gpkrq" Dec 01 08:57:22 crc kubenswrapper[4873]: I1201 08:57:22.925743 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34cbfd49-8920-4490-9f8a-7a6f2542d1c6-config\") pod \"dnsmasq-dns-666b6646f7-gpkrq\" (UID: \"34cbfd49-8920-4490-9f8a-7a6f2542d1c6\") " pod="openstack/dnsmasq-dns-666b6646f7-gpkrq" Dec 01 08:57:22 crc kubenswrapper[4873]: I1201 08:57:22.926489 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/34cbfd49-8920-4490-9f8a-7a6f2542d1c6-dns-svc\") pod \"dnsmasq-dns-666b6646f7-gpkrq\" (UID: \"34cbfd49-8920-4490-9f8a-7a6f2542d1c6\") " pod="openstack/dnsmasq-dns-666b6646f7-gpkrq" Dec 01 08:57:22 crc kubenswrapper[4873]: I1201 08:57:22.964618 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tzzd\" (UniqueName: \"kubernetes.io/projected/34cbfd49-8920-4490-9f8a-7a6f2542d1c6-kube-api-access-5tzzd\") pod \"dnsmasq-dns-666b6646f7-gpkrq\" (UID: \"34cbfd49-8920-4490-9f8a-7a6f2542d1c6\") " pod="openstack/dnsmasq-dns-666b6646f7-gpkrq" Dec 01 08:57:23 crc kubenswrapper[4873]: I1201 08:57:23.094572 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-gpkrq" Dec 01 08:57:23 crc kubenswrapper[4873]: I1201 08:57:23.144263 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-hssdh"] Dec 01 08:57:23 crc kubenswrapper[4873]: I1201 08:57:23.205195 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-zng4c"] Dec 01 08:57:23 crc kubenswrapper[4873]: I1201 08:57:23.207234 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-zng4c" Dec 01 08:57:23 crc kubenswrapper[4873]: I1201 08:57:23.233412 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d5588863-2c50-4860-ac45-4d226f59f4ab-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-zng4c\" (UID: \"d5588863-2c50-4860-ac45-4d226f59f4ab\") " pod="openstack/dnsmasq-dns-57d769cc4f-zng4c" Dec 01 08:57:23 crc kubenswrapper[4873]: I1201 08:57:23.233474 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smnfg\" (UniqueName: \"kubernetes.io/projected/d5588863-2c50-4860-ac45-4d226f59f4ab-kube-api-access-smnfg\") pod \"dnsmasq-dns-57d769cc4f-zng4c\" (UID: \"d5588863-2c50-4860-ac45-4d226f59f4ab\") " pod="openstack/dnsmasq-dns-57d769cc4f-zng4c" Dec 01 08:57:23 crc kubenswrapper[4873]: I1201 08:57:23.233571 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5588863-2c50-4860-ac45-4d226f59f4ab-config\") pod \"dnsmasq-dns-57d769cc4f-zng4c\" (UID: \"d5588863-2c50-4860-ac45-4d226f59f4ab\") " pod="openstack/dnsmasq-dns-57d769cc4f-zng4c" Dec 01 08:57:23 crc kubenswrapper[4873]: I1201 08:57:23.239752 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-zng4c"] Dec 01 08:57:23 crc kubenswrapper[4873]: I1201 08:57:23.334720 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d5588863-2c50-4860-ac45-4d226f59f4ab-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-zng4c\" (UID: \"d5588863-2c50-4860-ac45-4d226f59f4ab\") " pod="openstack/dnsmasq-dns-57d769cc4f-zng4c" Dec 01 08:57:23 crc kubenswrapper[4873]: I1201 08:57:23.334785 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smnfg\" (UniqueName: \"kubernetes.io/projected/d5588863-2c50-4860-ac45-4d226f59f4ab-kube-api-access-smnfg\") pod \"dnsmasq-dns-57d769cc4f-zng4c\" (UID: \"d5588863-2c50-4860-ac45-4d226f59f4ab\") " pod="openstack/dnsmasq-dns-57d769cc4f-zng4c" Dec 01 08:57:23 crc kubenswrapper[4873]: I1201 08:57:23.334866 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5588863-2c50-4860-ac45-4d226f59f4ab-config\") pod \"dnsmasq-dns-57d769cc4f-zng4c\" (UID: \"d5588863-2c50-4860-ac45-4d226f59f4ab\") " pod="openstack/dnsmasq-dns-57d769cc4f-zng4c" Dec 01 08:57:23 crc kubenswrapper[4873]: I1201 08:57:23.335758 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5588863-2c50-4860-ac45-4d226f59f4ab-config\") pod \"dnsmasq-dns-57d769cc4f-zng4c\" (UID: \"d5588863-2c50-4860-ac45-4d226f59f4ab\") " pod="openstack/dnsmasq-dns-57d769cc4f-zng4c" Dec 01 08:57:23 crc kubenswrapper[4873]: I1201 08:57:23.335795 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d5588863-2c50-4860-ac45-4d226f59f4ab-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-zng4c\" (UID: \"d5588863-2c50-4860-ac45-4d226f59f4ab\") " pod="openstack/dnsmasq-dns-57d769cc4f-zng4c" Dec 01 08:57:23 crc kubenswrapper[4873]: I1201 08:57:23.368003 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smnfg\" (UniqueName: \"kubernetes.io/projected/d5588863-2c50-4860-ac45-4d226f59f4ab-kube-api-access-smnfg\") pod \"dnsmasq-dns-57d769cc4f-zng4c\" (UID: \"d5588863-2c50-4860-ac45-4d226f59f4ab\") " pod="openstack/dnsmasq-dns-57d769cc4f-zng4c" Dec 01 08:57:23 crc kubenswrapper[4873]: I1201 08:57:23.539977 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-zng4c" Dec 01 08:57:23 crc kubenswrapper[4873]: I1201 08:57:23.798289 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-gpkrq"] Dec 01 08:57:23 crc kubenswrapper[4873]: I1201 08:57:23.835084 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-zng4c"] Dec 01 08:57:23 crc kubenswrapper[4873]: I1201 08:57:23.915124 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 08:57:23 crc kubenswrapper[4873]: I1201 08:57:23.918747 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 01 08:57:23 crc kubenswrapper[4873]: I1201 08:57:23.924830 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 01 08:57:23 crc kubenswrapper[4873]: I1201 08:57:23.925048 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 01 08:57:23 crc kubenswrapper[4873]: I1201 08:57:23.925201 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-td8xf" Dec 01 08:57:23 crc kubenswrapper[4873]: I1201 08:57:23.925279 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 01 08:57:23 crc kubenswrapper[4873]: I1201 08:57:23.925344 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 01 08:57:23 crc kubenswrapper[4873]: I1201 08:57:23.925747 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 01 08:57:23 crc kubenswrapper[4873]: I1201 08:57:23.928758 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 01 08:57:23 crc kubenswrapper[4873]: I1201 08:57:23.949700 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 08:57:23 crc kubenswrapper[4873]: I1201 08:57:23.951709 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6d27dec2-61e5-44ae-81df-8fee227ce3d8-config-data\") pod \"rabbitmq-server-0\" (UID: \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\") " pod="openstack/rabbitmq-server-0" Dec 01 08:57:23 crc kubenswrapper[4873]: I1201 08:57:23.951763 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6d27dec2-61e5-44ae-81df-8fee227ce3d8-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\") " pod="openstack/rabbitmq-server-0" Dec 01 08:57:23 crc kubenswrapper[4873]: I1201 08:57:23.951811 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6d27dec2-61e5-44ae-81df-8fee227ce3d8-pod-info\") pod \"rabbitmq-server-0\" (UID: \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\") " pod="openstack/rabbitmq-server-0" Dec 01 08:57:23 crc kubenswrapper[4873]: I1201 08:57:23.951906 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6d27dec2-61e5-44ae-81df-8fee227ce3d8-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\") " pod="openstack/rabbitmq-server-0" Dec 01 08:57:23 crc kubenswrapper[4873]: I1201 08:57:23.951933 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96hg6\" (UniqueName: \"kubernetes.io/projected/6d27dec2-61e5-44ae-81df-8fee227ce3d8-kube-api-access-96hg6\") pod \"rabbitmq-server-0\" (UID: \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\") " pod="openstack/rabbitmq-server-0" Dec 01 08:57:23 crc kubenswrapper[4873]: I1201 08:57:23.951973 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6d27dec2-61e5-44ae-81df-8fee227ce3d8-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\") " pod="openstack/rabbitmq-server-0" Dec 01 08:57:23 crc kubenswrapper[4873]: I1201 08:57:23.952026 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6d27dec2-61e5-44ae-81df-8fee227ce3d8-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\") " pod="openstack/rabbitmq-server-0" Dec 01 08:57:23 crc kubenswrapper[4873]: I1201 08:57:23.952196 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\") " pod="openstack/rabbitmq-server-0" Dec 01 08:57:23 crc kubenswrapper[4873]: I1201 08:57:23.952273 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6d27dec2-61e5-44ae-81df-8fee227ce3d8-server-conf\") pod \"rabbitmq-server-0\" (UID: \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\") " pod="openstack/rabbitmq-server-0" Dec 01 08:57:23 crc kubenswrapper[4873]: I1201 08:57:23.952422 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6d27dec2-61e5-44ae-81df-8fee227ce3d8-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\") " pod="openstack/rabbitmq-server-0" Dec 01 08:57:23 crc kubenswrapper[4873]: I1201 08:57:23.952501 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6d27dec2-61e5-44ae-81df-8fee227ce3d8-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\") " pod="openstack/rabbitmq-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.054745 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6d27dec2-61e5-44ae-81df-8fee227ce3d8-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\") " pod="openstack/rabbitmq-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.054853 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6d27dec2-61e5-44ae-81df-8fee227ce3d8-config-data\") pod \"rabbitmq-server-0\" (UID: \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\") " pod="openstack/rabbitmq-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.054883 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6d27dec2-61e5-44ae-81df-8fee227ce3d8-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\") " pod="openstack/rabbitmq-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.054923 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6d27dec2-61e5-44ae-81df-8fee227ce3d8-pod-info\") pod \"rabbitmq-server-0\" (UID: \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\") " pod="openstack/rabbitmq-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.054963 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6d27dec2-61e5-44ae-81df-8fee227ce3d8-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\") " pod="openstack/rabbitmq-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.054988 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96hg6\" (UniqueName: \"kubernetes.io/projected/6d27dec2-61e5-44ae-81df-8fee227ce3d8-kube-api-access-96hg6\") pod \"rabbitmq-server-0\" (UID: \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\") " pod="openstack/rabbitmq-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.055034 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6d27dec2-61e5-44ae-81df-8fee227ce3d8-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\") " pod="openstack/rabbitmq-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.055063 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6d27dec2-61e5-44ae-81df-8fee227ce3d8-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\") " pod="openstack/rabbitmq-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.055124 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\") " pod="openstack/rabbitmq-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.055365 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6d27dec2-61e5-44ae-81df-8fee227ce3d8-server-conf\") pod \"rabbitmq-server-0\" (UID: \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\") " pod="openstack/rabbitmq-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.055400 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6d27dec2-61e5-44ae-81df-8fee227ce3d8-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\") " pod="openstack/rabbitmq-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.055818 4873 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.058923 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6d27dec2-61e5-44ae-81df-8fee227ce3d8-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\") " pod="openstack/rabbitmq-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.059393 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6d27dec2-61e5-44ae-81df-8fee227ce3d8-config-data\") pod \"rabbitmq-server-0\" (UID: \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\") " pod="openstack/rabbitmq-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.059465 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6d27dec2-61e5-44ae-81df-8fee227ce3d8-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\") " pod="openstack/rabbitmq-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.061300 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6d27dec2-61e5-44ae-81df-8fee227ce3d8-server-conf\") pod \"rabbitmq-server-0\" (UID: \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\") " pod="openstack/rabbitmq-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.062643 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6d27dec2-61e5-44ae-81df-8fee227ce3d8-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\") " pod="openstack/rabbitmq-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.063577 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6d27dec2-61e5-44ae-81df-8fee227ce3d8-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\") " pod="openstack/rabbitmq-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.071815 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6d27dec2-61e5-44ae-81df-8fee227ce3d8-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\") " pod="openstack/rabbitmq-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.072214 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6d27dec2-61e5-44ae-81df-8fee227ce3d8-pod-info\") pod \"rabbitmq-server-0\" (UID: \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\") " pod="openstack/rabbitmq-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.073174 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6d27dec2-61e5-44ae-81df-8fee227ce3d8-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\") " pod="openstack/rabbitmq-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.078353 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96hg6\" (UniqueName: \"kubernetes.io/projected/6d27dec2-61e5-44ae-81df-8fee227ce3d8-kube-api-access-96hg6\") pod \"rabbitmq-server-0\" (UID: \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\") " pod="openstack/rabbitmq-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.089758 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\") " pod="openstack/rabbitmq-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.251859 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.400623 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.403027 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.405833 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.407000 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.407510 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.407655 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.407716 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.408360 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.416582 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-zgbnc" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.416891 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.468786 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/29423166-c350-44f8-97ff-adb8b2b99165-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"29423166-c350-44f8-97ff-adb8b2b99165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.468845 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/29423166-c350-44f8-97ff-adb8b2b99165-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"29423166-c350-44f8-97ff-adb8b2b99165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.468879 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/29423166-c350-44f8-97ff-adb8b2b99165-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"29423166-c350-44f8-97ff-adb8b2b99165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.468949 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"29423166-c350-44f8-97ff-adb8b2b99165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.468968 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/29423166-c350-44f8-97ff-adb8b2b99165-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"29423166-c350-44f8-97ff-adb8b2b99165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.469027 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qzmb\" (UniqueName: \"kubernetes.io/projected/29423166-c350-44f8-97ff-adb8b2b99165-kube-api-access-7qzmb\") pod \"rabbitmq-cell1-server-0\" (UID: \"29423166-c350-44f8-97ff-adb8b2b99165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.469056 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/29423166-c350-44f8-97ff-adb8b2b99165-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"29423166-c350-44f8-97ff-adb8b2b99165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.469096 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/29423166-c350-44f8-97ff-adb8b2b99165-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"29423166-c350-44f8-97ff-adb8b2b99165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.469113 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/29423166-c350-44f8-97ff-adb8b2b99165-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"29423166-c350-44f8-97ff-adb8b2b99165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.469130 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/29423166-c350-44f8-97ff-adb8b2b99165-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"29423166-c350-44f8-97ff-adb8b2b99165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.469150 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/29423166-c350-44f8-97ff-adb8b2b99165-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"29423166-c350-44f8-97ff-adb8b2b99165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.571172 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/29423166-c350-44f8-97ff-adb8b2b99165-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"29423166-c350-44f8-97ff-adb8b2b99165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.577405 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/29423166-c350-44f8-97ff-adb8b2b99165-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"29423166-c350-44f8-97ff-adb8b2b99165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.571234 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/29423166-c350-44f8-97ff-adb8b2b99165-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"29423166-c350-44f8-97ff-adb8b2b99165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.579899 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/29423166-c350-44f8-97ff-adb8b2b99165-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"29423166-c350-44f8-97ff-adb8b2b99165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.580081 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"29423166-c350-44f8-97ff-adb8b2b99165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.580146 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/29423166-c350-44f8-97ff-adb8b2b99165-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"29423166-c350-44f8-97ff-adb8b2b99165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.580227 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qzmb\" (UniqueName: \"kubernetes.io/projected/29423166-c350-44f8-97ff-adb8b2b99165-kube-api-access-7qzmb\") pod \"rabbitmq-cell1-server-0\" (UID: \"29423166-c350-44f8-97ff-adb8b2b99165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.580281 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/29423166-c350-44f8-97ff-adb8b2b99165-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"29423166-c350-44f8-97ff-adb8b2b99165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.580435 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/29423166-c350-44f8-97ff-adb8b2b99165-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"29423166-c350-44f8-97ff-adb8b2b99165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.580467 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/29423166-c350-44f8-97ff-adb8b2b99165-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"29423166-c350-44f8-97ff-adb8b2b99165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.580511 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/29423166-c350-44f8-97ff-adb8b2b99165-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"29423166-c350-44f8-97ff-adb8b2b99165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.580576 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/29423166-c350-44f8-97ff-adb8b2b99165-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"29423166-c350-44f8-97ff-adb8b2b99165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.584719 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/29423166-c350-44f8-97ff-adb8b2b99165-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"29423166-c350-44f8-97ff-adb8b2b99165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.585260 4873 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"29423166-c350-44f8-97ff-adb8b2b99165\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.585744 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/29423166-c350-44f8-97ff-adb8b2b99165-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"29423166-c350-44f8-97ff-adb8b2b99165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.586075 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/29423166-c350-44f8-97ff-adb8b2b99165-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"29423166-c350-44f8-97ff-adb8b2b99165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.587170 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/29423166-c350-44f8-97ff-adb8b2b99165-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"29423166-c350-44f8-97ff-adb8b2b99165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.588983 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/29423166-c350-44f8-97ff-adb8b2b99165-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"29423166-c350-44f8-97ff-adb8b2b99165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.590430 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/29423166-c350-44f8-97ff-adb8b2b99165-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"29423166-c350-44f8-97ff-adb8b2b99165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.592852 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/29423166-c350-44f8-97ff-adb8b2b99165-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"29423166-c350-44f8-97ff-adb8b2b99165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.604766 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/29423166-c350-44f8-97ff-adb8b2b99165-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"29423166-c350-44f8-97ff-adb8b2b99165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.630644 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qzmb\" (UniqueName: \"kubernetes.io/projected/29423166-c350-44f8-97ff-adb8b2b99165-kube-api-access-7qzmb\") pod \"rabbitmq-cell1-server-0\" (UID: \"29423166-c350-44f8-97ff-adb8b2b99165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.633978 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"29423166-c350-44f8-97ff-adb8b2b99165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.709697 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-gpkrq" event={"ID":"34cbfd49-8920-4490-9f8a-7a6f2542d1c6","Type":"ContainerStarted","Data":"33f8d169979df01a49e2cc720645b8540426da2ba6056033c1226b45556d345c"} Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.711189 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-zng4c" event={"ID":"d5588863-2c50-4860-ac45-4d226f59f4ab","Type":"ContainerStarted","Data":"7cbaa3ed8027f8a8dd10a930e57faa3ae56cff6b435a1f8f62351abfdf311464"} Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.764562 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:57:24 crc kubenswrapper[4873]: I1201 08:57:24.834276 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 08:57:24 crc kubenswrapper[4873]: W1201 08:57:24.838791 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6d27dec2_61e5_44ae_81df_8fee227ce3d8.slice/crio-0019bd449173284d5a8f1175b4ae0829a6e3e5bf457e17b6e629cfd3709c7620 WatchSource:0}: Error finding container 0019bd449173284d5a8f1175b4ae0829a6e3e5bf457e17b6e629cfd3709c7620: Status 404 returned error can't find the container with id 0019bd449173284d5a8f1175b4ae0829a6e3e5bf457e17b6e629cfd3709c7620 Dec 01 08:57:25 crc kubenswrapper[4873]: I1201 08:57:25.074756 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 08:57:25 crc kubenswrapper[4873]: I1201 08:57:25.534453 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 01 08:57:25 crc kubenswrapper[4873]: I1201 08:57:25.536746 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 01 08:57:25 crc kubenswrapper[4873]: I1201 08:57:25.544192 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 01 08:57:25 crc kubenswrapper[4873]: I1201 08:57:25.544635 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 01 08:57:25 crc kubenswrapper[4873]: I1201 08:57:25.550136 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-94jsr" Dec 01 08:57:25 crc kubenswrapper[4873]: I1201 08:57:25.550386 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 01 08:57:25 crc kubenswrapper[4873]: I1201 08:57:25.550587 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 01 08:57:25 crc kubenswrapper[4873]: I1201 08:57:25.556252 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 01 08:57:25 crc kubenswrapper[4873]: I1201 08:57:25.707069 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/250cbd69-c69a-475c-8973-bc98caf4b264-config-data-default\") pod \"openstack-galera-0\" (UID: \"250cbd69-c69a-475c-8973-bc98caf4b264\") " pod="openstack/openstack-galera-0" Dec 01 08:57:25 crc kubenswrapper[4873]: I1201 08:57:25.707481 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"250cbd69-c69a-475c-8973-bc98caf4b264\") " pod="openstack/openstack-galera-0" Dec 01 08:57:25 crc kubenswrapper[4873]: I1201 08:57:25.707538 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/250cbd69-c69a-475c-8973-bc98caf4b264-operator-scripts\") pod \"openstack-galera-0\" (UID: \"250cbd69-c69a-475c-8973-bc98caf4b264\") " pod="openstack/openstack-galera-0" Dec 01 08:57:25 crc kubenswrapper[4873]: I1201 08:57:25.707589 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/250cbd69-c69a-475c-8973-bc98caf4b264-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"250cbd69-c69a-475c-8973-bc98caf4b264\") " pod="openstack/openstack-galera-0" Dec 01 08:57:25 crc kubenswrapper[4873]: I1201 08:57:25.707613 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/250cbd69-c69a-475c-8973-bc98caf4b264-config-data-generated\") pod \"openstack-galera-0\" (UID: \"250cbd69-c69a-475c-8973-bc98caf4b264\") " pod="openstack/openstack-galera-0" Dec 01 08:57:25 crc kubenswrapper[4873]: I1201 08:57:25.707643 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrc8r\" (UniqueName: \"kubernetes.io/projected/250cbd69-c69a-475c-8973-bc98caf4b264-kube-api-access-vrc8r\") pod \"openstack-galera-0\" (UID: \"250cbd69-c69a-475c-8973-bc98caf4b264\") " pod="openstack/openstack-galera-0" Dec 01 08:57:25 crc kubenswrapper[4873]: I1201 08:57:25.707687 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/250cbd69-c69a-475c-8973-bc98caf4b264-kolla-config\") pod \"openstack-galera-0\" (UID: \"250cbd69-c69a-475c-8973-bc98caf4b264\") " pod="openstack/openstack-galera-0" Dec 01 08:57:25 crc kubenswrapper[4873]: I1201 08:57:25.707718 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/250cbd69-c69a-475c-8973-bc98caf4b264-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"250cbd69-c69a-475c-8973-bc98caf4b264\") " pod="openstack/openstack-galera-0" Dec 01 08:57:25 crc kubenswrapper[4873]: I1201 08:57:25.722531 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6d27dec2-61e5-44ae-81df-8fee227ce3d8","Type":"ContainerStarted","Data":"0019bd449173284d5a8f1175b4ae0829a6e3e5bf457e17b6e629cfd3709c7620"} Dec 01 08:57:25 crc kubenswrapper[4873]: I1201 08:57:25.809591 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/250cbd69-c69a-475c-8973-bc98caf4b264-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"250cbd69-c69a-475c-8973-bc98caf4b264\") " pod="openstack/openstack-galera-0" Dec 01 08:57:25 crc kubenswrapper[4873]: I1201 08:57:25.809667 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/250cbd69-c69a-475c-8973-bc98caf4b264-config-data-generated\") pod \"openstack-galera-0\" (UID: \"250cbd69-c69a-475c-8973-bc98caf4b264\") " pod="openstack/openstack-galera-0" Dec 01 08:57:25 crc kubenswrapper[4873]: I1201 08:57:25.809727 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrc8r\" (UniqueName: \"kubernetes.io/projected/250cbd69-c69a-475c-8973-bc98caf4b264-kube-api-access-vrc8r\") pod \"openstack-galera-0\" (UID: \"250cbd69-c69a-475c-8973-bc98caf4b264\") " pod="openstack/openstack-galera-0" Dec 01 08:57:25 crc kubenswrapper[4873]: I1201 08:57:25.809774 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/250cbd69-c69a-475c-8973-bc98caf4b264-kolla-config\") pod \"openstack-galera-0\" (UID: \"250cbd69-c69a-475c-8973-bc98caf4b264\") " pod="openstack/openstack-galera-0" Dec 01 08:57:25 crc kubenswrapper[4873]: I1201 08:57:25.809829 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/250cbd69-c69a-475c-8973-bc98caf4b264-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"250cbd69-c69a-475c-8973-bc98caf4b264\") " pod="openstack/openstack-galera-0" Dec 01 08:57:25 crc kubenswrapper[4873]: I1201 08:57:25.810660 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/250cbd69-c69a-475c-8973-bc98caf4b264-config-data-generated\") pod \"openstack-galera-0\" (UID: \"250cbd69-c69a-475c-8973-bc98caf4b264\") " pod="openstack/openstack-galera-0" Dec 01 08:57:25 crc kubenswrapper[4873]: I1201 08:57:25.810754 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/250cbd69-c69a-475c-8973-bc98caf4b264-config-data-default\") pod \"openstack-galera-0\" (UID: \"250cbd69-c69a-475c-8973-bc98caf4b264\") " pod="openstack/openstack-galera-0" Dec 01 08:57:25 crc kubenswrapper[4873]: I1201 08:57:25.810834 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"250cbd69-c69a-475c-8973-bc98caf4b264\") " pod="openstack/openstack-galera-0" Dec 01 08:57:25 crc kubenswrapper[4873]: I1201 08:57:25.810892 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/250cbd69-c69a-475c-8973-bc98caf4b264-operator-scripts\") pod \"openstack-galera-0\" (UID: \"250cbd69-c69a-475c-8973-bc98caf4b264\") " pod="openstack/openstack-galera-0" Dec 01 08:57:25 crc kubenswrapper[4873]: I1201 08:57:25.811872 4873 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"250cbd69-c69a-475c-8973-bc98caf4b264\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/openstack-galera-0" Dec 01 08:57:25 crc kubenswrapper[4873]: I1201 08:57:25.814379 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/250cbd69-c69a-475c-8973-bc98caf4b264-operator-scripts\") pod \"openstack-galera-0\" (UID: \"250cbd69-c69a-475c-8973-bc98caf4b264\") " pod="openstack/openstack-galera-0" Dec 01 08:57:25 crc kubenswrapper[4873]: I1201 08:57:25.815860 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/250cbd69-c69a-475c-8973-bc98caf4b264-config-data-default\") pod \"openstack-galera-0\" (UID: \"250cbd69-c69a-475c-8973-bc98caf4b264\") " pod="openstack/openstack-galera-0" Dec 01 08:57:25 crc kubenswrapper[4873]: I1201 08:57:25.818203 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/250cbd69-c69a-475c-8973-bc98caf4b264-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"250cbd69-c69a-475c-8973-bc98caf4b264\") " pod="openstack/openstack-galera-0" Dec 01 08:57:25 crc kubenswrapper[4873]: I1201 08:57:25.818421 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/250cbd69-c69a-475c-8973-bc98caf4b264-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"250cbd69-c69a-475c-8973-bc98caf4b264\") " pod="openstack/openstack-galera-0" Dec 01 08:57:25 crc kubenswrapper[4873]: I1201 08:57:25.820666 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/250cbd69-c69a-475c-8973-bc98caf4b264-kolla-config\") pod \"openstack-galera-0\" (UID: \"250cbd69-c69a-475c-8973-bc98caf4b264\") " pod="openstack/openstack-galera-0" Dec 01 08:57:25 crc kubenswrapper[4873]: I1201 08:57:25.846565 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"250cbd69-c69a-475c-8973-bc98caf4b264\") " pod="openstack/openstack-galera-0" Dec 01 08:57:25 crc kubenswrapper[4873]: I1201 08:57:25.873118 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrc8r\" (UniqueName: \"kubernetes.io/projected/250cbd69-c69a-475c-8973-bc98caf4b264-kube-api-access-vrc8r\") pod \"openstack-galera-0\" (UID: \"250cbd69-c69a-475c-8973-bc98caf4b264\") " pod="openstack/openstack-galera-0" Dec 01 08:57:26 crc kubenswrapper[4873]: I1201 08:57:26.165061 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 01 08:57:26 crc kubenswrapper[4873]: I1201 08:57:26.939512 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 01 08:57:26 crc kubenswrapper[4873]: I1201 08:57:26.941559 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 01 08:57:26 crc kubenswrapper[4873]: I1201 08:57:26.945452 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 01 08:57:26 crc kubenswrapper[4873]: I1201 08:57:26.945492 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-wdhbn" Dec 01 08:57:26 crc kubenswrapper[4873]: I1201 08:57:26.945747 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 01 08:57:26 crc kubenswrapper[4873]: I1201 08:57:26.945872 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 01 08:57:26 crc kubenswrapper[4873]: I1201 08:57:26.962148 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.040736 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"714ffca1-ee33-4f92-a55e-29e2edf84eff\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.040823 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/714ffca1-ee33-4f92-a55e-29e2edf84eff-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"714ffca1-ee33-4f92-a55e-29e2edf84eff\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.040859 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/714ffca1-ee33-4f92-a55e-29e2edf84eff-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"714ffca1-ee33-4f92-a55e-29e2edf84eff\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.040905 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/714ffca1-ee33-4f92-a55e-29e2edf84eff-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"714ffca1-ee33-4f92-a55e-29e2edf84eff\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.040943 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxtvt\" (UniqueName: \"kubernetes.io/projected/714ffca1-ee33-4f92-a55e-29e2edf84eff-kube-api-access-mxtvt\") pod \"openstack-cell1-galera-0\" (UID: \"714ffca1-ee33-4f92-a55e-29e2edf84eff\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.040989 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/714ffca1-ee33-4f92-a55e-29e2edf84eff-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"714ffca1-ee33-4f92-a55e-29e2edf84eff\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.041178 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/714ffca1-ee33-4f92-a55e-29e2edf84eff-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"714ffca1-ee33-4f92-a55e-29e2edf84eff\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.041399 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/714ffca1-ee33-4f92-a55e-29e2edf84eff-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"714ffca1-ee33-4f92-a55e-29e2edf84eff\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.149494 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/714ffca1-ee33-4f92-a55e-29e2edf84eff-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"714ffca1-ee33-4f92-a55e-29e2edf84eff\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.149567 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxtvt\" (UniqueName: \"kubernetes.io/projected/714ffca1-ee33-4f92-a55e-29e2edf84eff-kube-api-access-mxtvt\") pod \"openstack-cell1-galera-0\" (UID: \"714ffca1-ee33-4f92-a55e-29e2edf84eff\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.149617 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/714ffca1-ee33-4f92-a55e-29e2edf84eff-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"714ffca1-ee33-4f92-a55e-29e2edf84eff\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.149669 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/714ffca1-ee33-4f92-a55e-29e2edf84eff-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"714ffca1-ee33-4f92-a55e-29e2edf84eff\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.149731 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/714ffca1-ee33-4f92-a55e-29e2edf84eff-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"714ffca1-ee33-4f92-a55e-29e2edf84eff\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.149791 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"714ffca1-ee33-4f92-a55e-29e2edf84eff\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.149818 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/714ffca1-ee33-4f92-a55e-29e2edf84eff-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"714ffca1-ee33-4f92-a55e-29e2edf84eff\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.149840 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/714ffca1-ee33-4f92-a55e-29e2edf84eff-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"714ffca1-ee33-4f92-a55e-29e2edf84eff\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.149975 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/714ffca1-ee33-4f92-a55e-29e2edf84eff-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"714ffca1-ee33-4f92-a55e-29e2edf84eff\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.150393 4873 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"714ffca1-ee33-4f92-a55e-29e2edf84eff\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/openstack-cell1-galera-0" Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.151167 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/714ffca1-ee33-4f92-a55e-29e2edf84eff-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"714ffca1-ee33-4f92-a55e-29e2edf84eff\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.151285 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/714ffca1-ee33-4f92-a55e-29e2edf84eff-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"714ffca1-ee33-4f92-a55e-29e2edf84eff\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.151764 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/714ffca1-ee33-4f92-a55e-29e2edf84eff-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"714ffca1-ee33-4f92-a55e-29e2edf84eff\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.171849 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/714ffca1-ee33-4f92-a55e-29e2edf84eff-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"714ffca1-ee33-4f92-a55e-29e2edf84eff\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.189230 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxtvt\" (UniqueName: \"kubernetes.io/projected/714ffca1-ee33-4f92-a55e-29e2edf84eff-kube-api-access-mxtvt\") pod \"openstack-cell1-galera-0\" (UID: \"714ffca1-ee33-4f92-a55e-29e2edf84eff\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.208103 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/714ffca1-ee33-4f92-a55e-29e2edf84eff-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"714ffca1-ee33-4f92-a55e-29e2edf84eff\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.228310 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"714ffca1-ee33-4f92-a55e-29e2edf84eff\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.274578 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.323919 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.325264 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.329297 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.329608 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-882lq" Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.329977 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.354372 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.455295 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e6c09a7-97f9-4048-b5a9-b3df685eacbc-memcached-tls-certs\") pod \"memcached-0\" (UID: \"2e6c09a7-97f9-4048-b5a9-b3df685eacbc\") " pod="openstack/memcached-0" Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.455371 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e6c09a7-97f9-4048-b5a9-b3df685eacbc-combined-ca-bundle\") pod \"memcached-0\" (UID: \"2e6c09a7-97f9-4048-b5a9-b3df685eacbc\") " pod="openstack/memcached-0" Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.455430 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2e6c09a7-97f9-4048-b5a9-b3df685eacbc-config-data\") pod \"memcached-0\" (UID: \"2e6c09a7-97f9-4048-b5a9-b3df685eacbc\") " pod="openstack/memcached-0" Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.455453 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fb9kn\" (UniqueName: \"kubernetes.io/projected/2e6c09a7-97f9-4048-b5a9-b3df685eacbc-kube-api-access-fb9kn\") pod \"memcached-0\" (UID: \"2e6c09a7-97f9-4048-b5a9-b3df685eacbc\") " pod="openstack/memcached-0" Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.455499 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2e6c09a7-97f9-4048-b5a9-b3df685eacbc-kolla-config\") pod \"memcached-0\" (UID: \"2e6c09a7-97f9-4048-b5a9-b3df685eacbc\") " pod="openstack/memcached-0" Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.557346 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e6c09a7-97f9-4048-b5a9-b3df685eacbc-memcached-tls-certs\") pod \"memcached-0\" (UID: \"2e6c09a7-97f9-4048-b5a9-b3df685eacbc\") " pod="openstack/memcached-0" Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.557425 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e6c09a7-97f9-4048-b5a9-b3df685eacbc-combined-ca-bundle\") pod \"memcached-0\" (UID: \"2e6c09a7-97f9-4048-b5a9-b3df685eacbc\") " pod="openstack/memcached-0" Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.557479 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2e6c09a7-97f9-4048-b5a9-b3df685eacbc-config-data\") pod \"memcached-0\" (UID: \"2e6c09a7-97f9-4048-b5a9-b3df685eacbc\") " pod="openstack/memcached-0" Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.557512 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fb9kn\" (UniqueName: \"kubernetes.io/projected/2e6c09a7-97f9-4048-b5a9-b3df685eacbc-kube-api-access-fb9kn\") pod \"memcached-0\" (UID: \"2e6c09a7-97f9-4048-b5a9-b3df685eacbc\") " pod="openstack/memcached-0" Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.557565 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2e6c09a7-97f9-4048-b5a9-b3df685eacbc-kolla-config\") pod \"memcached-0\" (UID: \"2e6c09a7-97f9-4048-b5a9-b3df685eacbc\") " pod="openstack/memcached-0" Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.558483 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2e6c09a7-97f9-4048-b5a9-b3df685eacbc-config-data\") pod \"memcached-0\" (UID: \"2e6c09a7-97f9-4048-b5a9-b3df685eacbc\") " pod="openstack/memcached-0" Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.558933 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2e6c09a7-97f9-4048-b5a9-b3df685eacbc-kolla-config\") pod \"memcached-0\" (UID: \"2e6c09a7-97f9-4048-b5a9-b3df685eacbc\") " pod="openstack/memcached-0" Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.561862 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e6c09a7-97f9-4048-b5a9-b3df685eacbc-memcached-tls-certs\") pod \"memcached-0\" (UID: \"2e6c09a7-97f9-4048-b5a9-b3df685eacbc\") " pod="openstack/memcached-0" Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.562555 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e6c09a7-97f9-4048-b5a9-b3df685eacbc-combined-ca-bundle\") pod \"memcached-0\" (UID: \"2e6c09a7-97f9-4048-b5a9-b3df685eacbc\") " pod="openstack/memcached-0" Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.580942 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fb9kn\" (UniqueName: \"kubernetes.io/projected/2e6c09a7-97f9-4048-b5a9-b3df685eacbc-kube-api-access-fb9kn\") pod \"memcached-0\" (UID: \"2e6c09a7-97f9-4048-b5a9-b3df685eacbc\") " pod="openstack/memcached-0" Dec 01 08:57:27 crc kubenswrapper[4873]: I1201 08:57:27.645545 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 01 08:57:28 crc kubenswrapper[4873]: W1201 08:57:28.124915 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod29423166_c350_44f8_97ff_adb8b2b99165.slice/crio-4db2a1c7ca11408f44768168a39319eac9b7db6c879e07052042c18efb0f7c98 WatchSource:0}: Error finding container 4db2a1c7ca11408f44768168a39319eac9b7db6c879e07052042c18efb0f7c98: Status 404 returned error can't find the container with id 4db2a1c7ca11408f44768168a39319eac9b7db6c879e07052042c18efb0f7c98 Dec 01 08:57:28 crc kubenswrapper[4873]: I1201 08:57:28.766491 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"29423166-c350-44f8-97ff-adb8b2b99165","Type":"ContainerStarted","Data":"4db2a1c7ca11408f44768168a39319eac9b7db6c879e07052042c18efb0f7c98"} Dec 01 08:57:28 crc kubenswrapper[4873]: I1201 08:57:28.983836 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 08:57:28 crc kubenswrapper[4873]: I1201 08:57:28.985094 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 08:57:28 crc kubenswrapper[4873]: I1201 08:57:28.988267 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-sfvh6" Dec 01 08:57:28 crc kubenswrapper[4873]: I1201 08:57:28.996000 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 08:57:29 crc kubenswrapper[4873]: I1201 08:57:29.127528 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tr5ff\" (UniqueName: \"kubernetes.io/projected/c0ae1a05-ad24-4459-9a9c-1f4993efea08-kube-api-access-tr5ff\") pod \"kube-state-metrics-0\" (UID: \"c0ae1a05-ad24-4459-9a9c-1f4993efea08\") " pod="openstack/kube-state-metrics-0" Dec 01 08:57:29 crc kubenswrapper[4873]: I1201 08:57:29.229633 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tr5ff\" (UniqueName: \"kubernetes.io/projected/c0ae1a05-ad24-4459-9a9c-1f4993efea08-kube-api-access-tr5ff\") pod \"kube-state-metrics-0\" (UID: \"c0ae1a05-ad24-4459-9a9c-1f4993efea08\") " pod="openstack/kube-state-metrics-0" Dec 01 08:57:29 crc kubenswrapper[4873]: I1201 08:57:29.248390 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tr5ff\" (UniqueName: \"kubernetes.io/projected/c0ae1a05-ad24-4459-9a9c-1f4993efea08-kube-api-access-tr5ff\") pod \"kube-state-metrics-0\" (UID: \"c0ae1a05-ad24-4459-9a9c-1f4993efea08\") " pod="openstack/kube-state-metrics-0" Dec 01 08:57:29 crc kubenswrapper[4873]: I1201 08:57:29.304108 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 08:57:32 crc kubenswrapper[4873]: I1201 08:57:32.948642 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-w8v6j"] Dec 01 08:57:32 crc kubenswrapper[4873]: I1201 08:57:32.954918 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-w8v6j" Dec 01 08:57:32 crc kubenswrapper[4873]: I1201 08:57:32.958630 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 01 08:57:32 crc kubenswrapper[4873]: I1201 08:57:32.958660 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-psql2" Dec 01 08:57:32 crc kubenswrapper[4873]: I1201 08:57:32.958863 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 01 08:57:32 crc kubenswrapper[4873]: I1201 08:57:32.984448 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b79364c5-00b2-4ab8-b2f0-aff17ed902b4-combined-ca-bundle\") pod \"ovn-controller-w8v6j\" (UID: \"b79364c5-00b2-4ab8-b2f0-aff17ed902b4\") " pod="openstack/ovn-controller-w8v6j" Dec 01 08:57:32 crc kubenswrapper[4873]: I1201 08:57:32.984881 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b79364c5-00b2-4ab8-b2f0-aff17ed902b4-var-run\") pod \"ovn-controller-w8v6j\" (UID: \"b79364c5-00b2-4ab8-b2f0-aff17ed902b4\") " pod="openstack/ovn-controller-w8v6j" Dec 01 08:57:32 crc kubenswrapper[4873]: I1201 08:57:32.985218 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgskk\" (UniqueName: \"kubernetes.io/projected/b79364c5-00b2-4ab8-b2f0-aff17ed902b4-kube-api-access-hgskk\") pod \"ovn-controller-w8v6j\" (UID: \"b79364c5-00b2-4ab8-b2f0-aff17ed902b4\") " pod="openstack/ovn-controller-w8v6j" Dec 01 08:57:32 crc kubenswrapper[4873]: I1201 08:57:32.985472 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b79364c5-00b2-4ab8-b2f0-aff17ed902b4-var-run-ovn\") pod \"ovn-controller-w8v6j\" (UID: \"b79364c5-00b2-4ab8-b2f0-aff17ed902b4\") " pod="openstack/ovn-controller-w8v6j" Dec 01 08:57:32 crc kubenswrapper[4873]: I1201 08:57:32.985542 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/b79364c5-00b2-4ab8-b2f0-aff17ed902b4-ovn-controller-tls-certs\") pod \"ovn-controller-w8v6j\" (UID: \"b79364c5-00b2-4ab8-b2f0-aff17ed902b4\") " pod="openstack/ovn-controller-w8v6j" Dec 01 08:57:32 crc kubenswrapper[4873]: I1201 08:57:32.985565 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b79364c5-00b2-4ab8-b2f0-aff17ed902b4-var-log-ovn\") pod \"ovn-controller-w8v6j\" (UID: \"b79364c5-00b2-4ab8-b2f0-aff17ed902b4\") " pod="openstack/ovn-controller-w8v6j" Dec 01 08:57:32 crc kubenswrapper[4873]: I1201 08:57:32.985582 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b79364c5-00b2-4ab8-b2f0-aff17ed902b4-scripts\") pod \"ovn-controller-w8v6j\" (UID: \"b79364c5-00b2-4ab8-b2f0-aff17ed902b4\") " pod="openstack/ovn-controller-w8v6j" Dec 01 08:57:33 crc kubenswrapper[4873]: I1201 08:57:33.002539 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-w8v6j"] Dec 01 08:57:33 crc kubenswrapper[4873]: I1201 08:57:33.035560 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-xd66n"] Dec 01 08:57:33 crc kubenswrapper[4873]: I1201 08:57:33.037871 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-xd66n" Dec 01 08:57:33 crc kubenswrapper[4873]: I1201 08:57:33.080877 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-xd66n"] Dec 01 08:57:33 crc kubenswrapper[4873]: I1201 08:57:33.087140 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b79364c5-00b2-4ab8-b2f0-aff17ed902b4-combined-ca-bundle\") pod \"ovn-controller-w8v6j\" (UID: \"b79364c5-00b2-4ab8-b2f0-aff17ed902b4\") " pod="openstack/ovn-controller-w8v6j" Dec 01 08:57:33 crc kubenswrapper[4873]: I1201 08:57:33.087197 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b79364c5-00b2-4ab8-b2f0-aff17ed902b4-var-run\") pod \"ovn-controller-w8v6j\" (UID: \"b79364c5-00b2-4ab8-b2f0-aff17ed902b4\") " pod="openstack/ovn-controller-w8v6j" Dec 01 08:57:33 crc kubenswrapper[4873]: I1201 08:57:33.087269 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgskk\" (UniqueName: \"kubernetes.io/projected/b79364c5-00b2-4ab8-b2f0-aff17ed902b4-kube-api-access-hgskk\") pod \"ovn-controller-w8v6j\" (UID: \"b79364c5-00b2-4ab8-b2f0-aff17ed902b4\") " pod="openstack/ovn-controller-w8v6j" Dec 01 08:57:33 crc kubenswrapper[4873]: I1201 08:57:33.087302 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/243fec55-bb19-4807-8b15-69584a922cf6-var-log\") pod \"ovn-controller-ovs-xd66n\" (UID: \"243fec55-bb19-4807-8b15-69584a922cf6\") " pod="openstack/ovn-controller-ovs-xd66n" Dec 01 08:57:33 crc kubenswrapper[4873]: I1201 08:57:33.087349 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmrtf\" (UniqueName: \"kubernetes.io/projected/243fec55-bb19-4807-8b15-69584a922cf6-kube-api-access-mmrtf\") pod \"ovn-controller-ovs-xd66n\" (UID: \"243fec55-bb19-4807-8b15-69584a922cf6\") " pod="openstack/ovn-controller-ovs-xd66n" Dec 01 08:57:33 crc kubenswrapper[4873]: I1201 08:57:33.087375 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/243fec55-bb19-4807-8b15-69584a922cf6-etc-ovs\") pod \"ovn-controller-ovs-xd66n\" (UID: \"243fec55-bb19-4807-8b15-69584a922cf6\") " pod="openstack/ovn-controller-ovs-xd66n" Dec 01 08:57:33 crc kubenswrapper[4873]: I1201 08:57:33.087396 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/243fec55-bb19-4807-8b15-69584a922cf6-var-lib\") pod \"ovn-controller-ovs-xd66n\" (UID: \"243fec55-bb19-4807-8b15-69584a922cf6\") " pod="openstack/ovn-controller-ovs-xd66n" Dec 01 08:57:33 crc kubenswrapper[4873]: I1201 08:57:33.087418 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b79364c5-00b2-4ab8-b2f0-aff17ed902b4-var-run-ovn\") pod \"ovn-controller-w8v6j\" (UID: \"b79364c5-00b2-4ab8-b2f0-aff17ed902b4\") " pod="openstack/ovn-controller-w8v6j" Dec 01 08:57:33 crc kubenswrapper[4873]: I1201 08:57:33.087695 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/b79364c5-00b2-4ab8-b2f0-aff17ed902b4-ovn-controller-tls-certs\") pod \"ovn-controller-w8v6j\" (UID: \"b79364c5-00b2-4ab8-b2f0-aff17ed902b4\") " pod="openstack/ovn-controller-w8v6j" Dec 01 08:57:33 crc kubenswrapper[4873]: I1201 08:57:33.087762 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b79364c5-00b2-4ab8-b2f0-aff17ed902b4-var-log-ovn\") pod \"ovn-controller-w8v6j\" (UID: \"b79364c5-00b2-4ab8-b2f0-aff17ed902b4\") " pod="openstack/ovn-controller-w8v6j" Dec 01 08:57:33 crc kubenswrapper[4873]: I1201 08:57:33.087786 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b79364c5-00b2-4ab8-b2f0-aff17ed902b4-scripts\") pod \"ovn-controller-w8v6j\" (UID: \"b79364c5-00b2-4ab8-b2f0-aff17ed902b4\") " pod="openstack/ovn-controller-w8v6j" Dec 01 08:57:33 crc kubenswrapper[4873]: I1201 08:57:33.087843 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/243fec55-bb19-4807-8b15-69584a922cf6-var-run\") pod \"ovn-controller-ovs-xd66n\" (UID: \"243fec55-bb19-4807-8b15-69584a922cf6\") " pod="openstack/ovn-controller-ovs-xd66n" Dec 01 08:57:33 crc kubenswrapper[4873]: I1201 08:57:33.087885 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/243fec55-bb19-4807-8b15-69584a922cf6-scripts\") pod \"ovn-controller-ovs-xd66n\" (UID: \"243fec55-bb19-4807-8b15-69584a922cf6\") " pod="openstack/ovn-controller-ovs-xd66n" Dec 01 08:57:33 crc kubenswrapper[4873]: I1201 08:57:33.087908 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b79364c5-00b2-4ab8-b2f0-aff17ed902b4-var-run\") pod \"ovn-controller-w8v6j\" (UID: \"b79364c5-00b2-4ab8-b2f0-aff17ed902b4\") " pod="openstack/ovn-controller-w8v6j" Dec 01 08:57:33 crc kubenswrapper[4873]: I1201 08:57:33.087975 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b79364c5-00b2-4ab8-b2f0-aff17ed902b4-var-run-ovn\") pod \"ovn-controller-w8v6j\" (UID: \"b79364c5-00b2-4ab8-b2f0-aff17ed902b4\") " pod="openstack/ovn-controller-w8v6j" Dec 01 08:57:33 crc kubenswrapper[4873]: I1201 08:57:33.088023 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b79364c5-00b2-4ab8-b2f0-aff17ed902b4-var-log-ovn\") pod \"ovn-controller-w8v6j\" (UID: \"b79364c5-00b2-4ab8-b2f0-aff17ed902b4\") " pod="openstack/ovn-controller-w8v6j" Dec 01 08:57:33 crc kubenswrapper[4873]: I1201 08:57:33.090773 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b79364c5-00b2-4ab8-b2f0-aff17ed902b4-scripts\") pod \"ovn-controller-w8v6j\" (UID: \"b79364c5-00b2-4ab8-b2f0-aff17ed902b4\") " pod="openstack/ovn-controller-w8v6j" Dec 01 08:57:33 crc kubenswrapper[4873]: I1201 08:57:33.102122 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/b79364c5-00b2-4ab8-b2f0-aff17ed902b4-ovn-controller-tls-certs\") pod \"ovn-controller-w8v6j\" (UID: \"b79364c5-00b2-4ab8-b2f0-aff17ed902b4\") " pod="openstack/ovn-controller-w8v6j" Dec 01 08:57:33 crc kubenswrapper[4873]: I1201 08:57:33.127464 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgskk\" (UniqueName: \"kubernetes.io/projected/b79364c5-00b2-4ab8-b2f0-aff17ed902b4-kube-api-access-hgskk\") pod \"ovn-controller-w8v6j\" (UID: \"b79364c5-00b2-4ab8-b2f0-aff17ed902b4\") " pod="openstack/ovn-controller-w8v6j" Dec 01 08:57:33 crc kubenswrapper[4873]: I1201 08:57:33.128634 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b79364c5-00b2-4ab8-b2f0-aff17ed902b4-combined-ca-bundle\") pod \"ovn-controller-w8v6j\" (UID: \"b79364c5-00b2-4ab8-b2f0-aff17ed902b4\") " pod="openstack/ovn-controller-w8v6j" Dec 01 08:57:33 crc kubenswrapper[4873]: I1201 08:57:33.190474 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/243fec55-bb19-4807-8b15-69584a922cf6-etc-ovs\") pod \"ovn-controller-ovs-xd66n\" (UID: \"243fec55-bb19-4807-8b15-69584a922cf6\") " pod="openstack/ovn-controller-ovs-xd66n" Dec 01 08:57:33 crc kubenswrapper[4873]: I1201 08:57:33.190545 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/243fec55-bb19-4807-8b15-69584a922cf6-var-lib\") pod \"ovn-controller-ovs-xd66n\" (UID: \"243fec55-bb19-4807-8b15-69584a922cf6\") " pod="openstack/ovn-controller-ovs-xd66n" Dec 01 08:57:33 crc kubenswrapper[4873]: I1201 08:57:33.190636 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/243fec55-bb19-4807-8b15-69584a922cf6-var-run\") pod \"ovn-controller-ovs-xd66n\" (UID: \"243fec55-bb19-4807-8b15-69584a922cf6\") " pod="openstack/ovn-controller-ovs-xd66n" Dec 01 08:57:33 crc kubenswrapper[4873]: I1201 08:57:33.190665 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/243fec55-bb19-4807-8b15-69584a922cf6-scripts\") pod \"ovn-controller-ovs-xd66n\" (UID: \"243fec55-bb19-4807-8b15-69584a922cf6\") " pod="openstack/ovn-controller-ovs-xd66n" Dec 01 08:57:33 crc kubenswrapper[4873]: I1201 08:57:33.190791 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/243fec55-bb19-4807-8b15-69584a922cf6-var-log\") pod \"ovn-controller-ovs-xd66n\" (UID: \"243fec55-bb19-4807-8b15-69584a922cf6\") " pod="openstack/ovn-controller-ovs-xd66n" Dec 01 08:57:33 crc kubenswrapper[4873]: I1201 08:57:33.190812 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/243fec55-bb19-4807-8b15-69584a922cf6-etc-ovs\") pod \"ovn-controller-ovs-xd66n\" (UID: \"243fec55-bb19-4807-8b15-69584a922cf6\") " pod="openstack/ovn-controller-ovs-xd66n" Dec 01 08:57:33 crc kubenswrapper[4873]: I1201 08:57:33.190865 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmrtf\" (UniqueName: \"kubernetes.io/projected/243fec55-bb19-4807-8b15-69584a922cf6-kube-api-access-mmrtf\") pod \"ovn-controller-ovs-xd66n\" (UID: \"243fec55-bb19-4807-8b15-69584a922cf6\") " pod="openstack/ovn-controller-ovs-xd66n" Dec 01 08:57:33 crc kubenswrapper[4873]: I1201 08:57:33.190907 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/243fec55-bb19-4807-8b15-69584a922cf6-var-run\") pod \"ovn-controller-ovs-xd66n\" (UID: \"243fec55-bb19-4807-8b15-69584a922cf6\") " pod="openstack/ovn-controller-ovs-xd66n" Dec 01 08:57:33 crc kubenswrapper[4873]: I1201 08:57:33.191097 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/243fec55-bb19-4807-8b15-69584a922cf6-var-lib\") pod \"ovn-controller-ovs-xd66n\" (UID: \"243fec55-bb19-4807-8b15-69584a922cf6\") " pod="openstack/ovn-controller-ovs-xd66n" Dec 01 08:57:33 crc kubenswrapper[4873]: I1201 08:57:33.191652 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/243fec55-bb19-4807-8b15-69584a922cf6-var-log\") pod \"ovn-controller-ovs-xd66n\" (UID: \"243fec55-bb19-4807-8b15-69584a922cf6\") " pod="openstack/ovn-controller-ovs-xd66n" Dec 01 08:57:33 crc kubenswrapper[4873]: I1201 08:57:33.194167 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/243fec55-bb19-4807-8b15-69584a922cf6-scripts\") pod \"ovn-controller-ovs-xd66n\" (UID: \"243fec55-bb19-4807-8b15-69584a922cf6\") " pod="openstack/ovn-controller-ovs-xd66n" Dec 01 08:57:33 crc kubenswrapper[4873]: I1201 08:57:33.218662 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmrtf\" (UniqueName: \"kubernetes.io/projected/243fec55-bb19-4807-8b15-69584a922cf6-kube-api-access-mmrtf\") pod \"ovn-controller-ovs-xd66n\" (UID: \"243fec55-bb19-4807-8b15-69584a922cf6\") " pod="openstack/ovn-controller-ovs-xd66n" Dec 01 08:57:33 crc kubenswrapper[4873]: I1201 08:57:33.281331 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-w8v6j" Dec 01 08:57:33 crc kubenswrapper[4873]: I1201 08:57:33.362388 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-xd66n" Dec 01 08:57:33 crc kubenswrapper[4873]: I1201 08:57:33.947089 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 01 08:57:33 crc kubenswrapper[4873]: I1201 08:57:33.949183 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 01 08:57:33 crc kubenswrapper[4873]: I1201 08:57:33.952391 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 01 08:57:33 crc kubenswrapper[4873]: I1201 08:57:33.952446 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 01 08:57:33 crc kubenswrapper[4873]: I1201 08:57:33.952573 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 01 08:57:33 crc kubenswrapper[4873]: I1201 08:57:33.952401 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 01 08:57:33 crc kubenswrapper[4873]: I1201 08:57:33.953688 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-x7xcp" Dec 01 08:57:33 crc kubenswrapper[4873]: I1201 08:57:33.968002 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 01 08:57:34 crc kubenswrapper[4873]: I1201 08:57:34.114495 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/21c27fff-80e6-4a74-aad2-5e772b811535-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"21c27fff-80e6-4a74-aad2-5e772b811535\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:57:34 crc kubenswrapper[4873]: I1201 08:57:34.114619 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/21c27fff-80e6-4a74-aad2-5e772b811535-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"21c27fff-80e6-4a74-aad2-5e772b811535\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:57:34 crc kubenswrapper[4873]: I1201 08:57:34.114652 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czsd7\" (UniqueName: \"kubernetes.io/projected/21c27fff-80e6-4a74-aad2-5e772b811535-kube-api-access-czsd7\") pod \"ovsdbserver-nb-0\" (UID: \"21c27fff-80e6-4a74-aad2-5e772b811535\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:57:34 crc kubenswrapper[4873]: I1201 08:57:34.114681 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-nb-0\" (UID: \"21c27fff-80e6-4a74-aad2-5e772b811535\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:57:34 crc kubenswrapper[4873]: I1201 08:57:34.114709 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21c27fff-80e6-4a74-aad2-5e772b811535-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"21c27fff-80e6-4a74-aad2-5e772b811535\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:57:34 crc kubenswrapper[4873]: I1201 08:57:34.114754 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/21c27fff-80e6-4a74-aad2-5e772b811535-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"21c27fff-80e6-4a74-aad2-5e772b811535\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:57:34 crc kubenswrapper[4873]: I1201 08:57:34.114798 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21c27fff-80e6-4a74-aad2-5e772b811535-config\") pod \"ovsdbserver-nb-0\" (UID: \"21c27fff-80e6-4a74-aad2-5e772b811535\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:57:34 crc kubenswrapper[4873]: I1201 08:57:34.114821 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/21c27fff-80e6-4a74-aad2-5e772b811535-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"21c27fff-80e6-4a74-aad2-5e772b811535\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:57:34 crc kubenswrapper[4873]: I1201 08:57:34.217813 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/21c27fff-80e6-4a74-aad2-5e772b811535-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"21c27fff-80e6-4a74-aad2-5e772b811535\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:57:34 crc kubenswrapper[4873]: I1201 08:57:34.219189 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czsd7\" (UniqueName: \"kubernetes.io/projected/21c27fff-80e6-4a74-aad2-5e772b811535-kube-api-access-czsd7\") pod \"ovsdbserver-nb-0\" (UID: \"21c27fff-80e6-4a74-aad2-5e772b811535\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:57:34 crc kubenswrapper[4873]: I1201 08:57:34.219633 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-nb-0\" (UID: \"21c27fff-80e6-4a74-aad2-5e772b811535\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:57:34 crc kubenswrapper[4873]: I1201 08:57:34.220081 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21c27fff-80e6-4a74-aad2-5e772b811535-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"21c27fff-80e6-4a74-aad2-5e772b811535\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:57:34 crc kubenswrapper[4873]: I1201 08:57:34.220677 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/21c27fff-80e6-4a74-aad2-5e772b811535-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"21c27fff-80e6-4a74-aad2-5e772b811535\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:57:34 crc kubenswrapper[4873]: I1201 08:57:34.220009 4873 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-nb-0\" (UID: \"21c27fff-80e6-4a74-aad2-5e772b811535\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/ovsdbserver-nb-0" Dec 01 08:57:34 crc kubenswrapper[4873]: I1201 08:57:34.220744 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21c27fff-80e6-4a74-aad2-5e772b811535-config\") pod \"ovsdbserver-nb-0\" (UID: \"21c27fff-80e6-4a74-aad2-5e772b811535\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:57:34 crc kubenswrapper[4873]: I1201 08:57:34.220775 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/21c27fff-80e6-4a74-aad2-5e772b811535-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"21c27fff-80e6-4a74-aad2-5e772b811535\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:57:34 crc kubenswrapper[4873]: I1201 08:57:34.219109 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/21c27fff-80e6-4a74-aad2-5e772b811535-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"21c27fff-80e6-4a74-aad2-5e772b811535\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:57:34 crc kubenswrapper[4873]: I1201 08:57:34.220863 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/21c27fff-80e6-4a74-aad2-5e772b811535-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"21c27fff-80e6-4a74-aad2-5e772b811535\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:57:34 crc kubenswrapper[4873]: I1201 08:57:34.221615 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/21c27fff-80e6-4a74-aad2-5e772b811535-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"21c27fff-80e6-4a74-aad2-5e772b811535\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:57:34 crc kubenswrapper[4873]: I1201 08:57:34.221982 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21c27fff-80e6-4a74-aad2-5e772b811535-config\") pod \"ovsdbserver-nb-0\" (UID: \"21c27fff-80e6-4a74-aad2-5e772b811535\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:57:34 crc kubenswrapper[4873]: I1201 08:57:34.224859 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21c27fff-80e6-4a74-aad2-5e772b811535-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"21c27fff-80e6-4a74-aad2-5e772b811535\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:57:34 crc kubenswrapper[4873]: I1201 08:57:34.226656 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/21c27fff-80e6-4a74-aad2-5e772b811535-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"21c27fff-80e6-4a74-aad2-5e772b811535\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:57:34 crc kubenswrapper[4873]: I1201 08:57:34.230199 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/21c27fff-80e6-4a74-aad2-5e772b811535-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"21c27fff-80e6-4a74-aad2-5e772b811535\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:57:34 crc kubenswrapper[4873]: I1201 08:57:34.236672 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czsd7\" (UniqueName: \"kubernetes.io/projected/21c27fff-80e6-4a74-aad2-5e772b811535-kube-api-access-czsd7\") pod \"ovsdbserver-nb-0\" (UID: \"21c27fff-80e6-4a74-aad2-5e772b811535\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:57:34 crc kubenswrapper[4873]: I1201 08:57:34.249121 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-nb-0\" (UID: \"21c27fff-80e6-4a74-aad2-5e772b811535\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:57:34 crc kubenswrapper[4873]: I1201 08:57:34.279436 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 01 08:57:36 crc kubenswrapper[4873]: I1201 08:57:36.178547 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 01 08:57:36 crc kubenswrapper[4873]: I1201 08:57:36.180131 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 01 08:57:36 crc kubenswrapper[4873]: I1201 08:57:36.188818 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 01 08:57:36 crc kubenswrapper[4873]: I1201 08:57:36.188965 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-wrghj" Dec 01 08:57:36 crc kubenswrapper[4873]: I1201 08:57:36.189174 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 01 08:57:36 crc kubenswrapper[4873]: I1201 08:57:36.191669 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 01 08:57:36 crc kubenswrapper[4873]: I1201 08:57:36.202775 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 01 08:57:36 crc kubenswrapper[4873]: I1201 08:57:36.362469 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbe0e07d-50a3-47c2-831d-69f24c0e838a-config\") pod \"ovsdbserver-sb-0\" (UID: \"dbe0e07d-50a3-47c2-831d-69f24c0e838a\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:57:36 crc kubenswrapper[4873]: I1201 08:57:36.363175 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dbe0e07d-50a3-47c2-831d-69f24c0e838a-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"dbe0e07d-50a3-47c2-831d-69f24c0e838a\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:57:36 crc kubenswrapper[4873]: I1201 08:57:36.363214 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnstb\" (UniqueName: \"kubernetes.io/projected/dbe0e07d-50a3-47c2-831d-69f24c0e838a-kube-api-access-dnstb\") pod \"ovsdbserver-sb-0\" (UID: \"dbe0e07d-50a3-47c2-831d-69f24c0e838a\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:57:36 crc kubenswrapper[4873]: I1201 08:57:36.363248 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbe0e07d-50a3-47c2-831d-69f24c0e838a-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"dbe0e07d-50a3-47c2-831d-69f24c0e838a\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:57:36 crc kubenswrapper[4873]: I1201 08:57:36.363299 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbe0e07d-50a3-47c2-831d-69f24c0e838a-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"dbe0e07d-50a3-47c2-831d-69f24c0e838a\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:57:36 crc kubenswrapper[4873]: I1201 08:57:36.363448 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"dbe0e07d-50a3-47c2-831d-69f24c0e838a\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:57:36 crc kubenswrapper[4873]: I1201 08:57:36.363758 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dbe0e07d-50a3-47c2-831d-69f24c0e838a-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"dbe0e07d-50a3-47c2-831d-69f24c0e838a\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:57:36 crc kubenswrapper[4873]: I1201 08:57:36.363810 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbe0e07d-50a3-47c2-831d-69f24c0e838a-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"dbe0e07d-50a3-47c2-831d-69f24c0e838a\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:57:36 crc kubenswrapper[4873]: I1201 08:57:36.466724 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbe0e07d-50a3-47c2-831d-69f24c0e838a-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"dbe0e07d-50a3-47c2-831d-69f24c0e838a\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:57:36 crc kubenswrapper[4873]: I1201 08:57:36.467060 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbe0e07d-50a3-47c2-831d-69f24c0e838a-config\") pod \"ovsdbserver-sb-0\" (UID: \"dbe0e07d-50a3-47c2-831d-69f24c0e838a\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:57:36 crc kubenswrapper[4873]: I1201 08:57:36.467149 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dbe0e07d-50a3-47c2-831d-69f24c0e838a-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"dbe0e07d-50a3-47c2-831d-69f24c0e838a\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:57:36 crc kubenswrapper[4873]: I1201 08:57:36.467282 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnstb\" (UniqueName: \"kubernetes.io/projected/dbe0e07d-50a3-47c2-831d-69f24c0e838a-kube-api-access-dnstb\") pod \"ovsdbserver-sb-0\" (UID: \"dbe0e07d-50a3-47c2-831d-69f24c0e838a\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:57:36 crc kubenswrapper[4873]: I1201 08:57:36.467341 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbe0e07d-50a3-47c2-831d-69f24c0e838a-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"dbe0e07d-50a3-47c2-831d-69f24c0e838a\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:57:36 crc kubenswrapper[4873]: I1201 08:57:36.467743 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbe0e07d-50a3-47c2-831d-69f24c0e838a-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"dbe0e07d-50a3-47c2-831d-69f24c0e838a\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:57:36 crc kubenswrapper[4873]: I1201 08:57:36.468322 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dbe0e07d-50a3-47c2-831d-69f24c0e838a-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"dbe0e07d-50a3-47c2-831d-69f24c0e838a\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:57:36 crc kubenswrapper[4873]: I1201 08:57:36.468423 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"dbe0e07d-50a3-47c2-831d-69f24c0e838a\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:57:36 crc kubenswrapper[4873]: I1201 08:57:36.468463 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dbe0e07d-50a3-47c2-831d-69f24c0e838a-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"dbe0e07d-50a3-47c2-831d-69f24c0e838a\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:57:36 crc kubenswrapper[4873]: I1201 08:57:36.468566 4873 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"dbe0e07d-50a3-47c2-831d-69f24c0e838a\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/ovsdbserver-sb-0" Dec 01 08:57:36 crc kubenswrapper[4873]: I1201 08:57:36.468751 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbe0e07d-50a3-47c2-831d-69f24c0e838a-config\") pod \"ovsdbserver-sb-0\" (UID: \"dbe0e07d-50a3-47c2-831d-69f24c0e838a\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:57:36 crc kubenswrapper[4873]: I1201 08:57:36.469837 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dbe0e07d-50a3-47c2-831d-69f24c0e838a-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"dbe0e07d-50a3-47c2-831d-69f24c0e838a\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:57:36 crc kubenswrapper[4873]: I1201 08:57:36.474447 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbe0e07d-50a3-47c2-831d-69f24c0e838a-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"dbe0e07d-50a3-47c2-831d-69f24c0e838a\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:57:36 crc kubenswrapper[4873]: I1201 08:57:36.474935 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbe0e07d-50a3-47c2-831d-69f24c0e838a-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"dbe0e07d-50a3-47c2-831d-69f24c0e838a\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:57:36 crc kubenswrapper[4873]: I1201 08:57:36.480454 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbe0e07d-50a3-47c2-831d-69f24c0e838a-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"dbe0e07d-50a3-47c2-831d-69f24c0e838a\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:57:36 crc kubenswrapper[4873]: I1201 08:57:36.488608 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnstb\" (UniqueName: \"kubernetes.io/projected/dbe0e07d-50a3-47c2-831d-69f24c0e838a-kube-api-access-dnstb\") pod \"ovsdbserver-sb-0\" (UID: \"dbe0e07d-50a3-47c2-831d-69f24c0e838a\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:57:36 crc kubenswrapper[4873]: I1201 08:57:36.729746 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"dbe0e07d-50a3-47c2-831d-69f24c0e838a\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:57:36 crc kubenswrapper[4873]: I1201 08:57:36.818007 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 01 08:57:48 crc kubenswrapper[4873]: E1201 08:57:48.626544 4873 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Dec 01 08:57:48 crc kubenswrapper[4873]: E1201 08:57:48.626974 4873 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-96hg6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(6d27dec2-61e5-44ae-81df-8fee227ce3d8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 08:57:48 crc kubenswrapper[4873]: E1201 08:57:48.628144 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="6d27dec2-61e5-44ae-81df-8fee227ce3d8" Dec 01 08:57:48 crc kubenswrapper[4873]: E1201 08:57:48.672516 4873 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Dec 01 08:57:48 crc kubenswrapper[4873]: E1201 08:57:48.673117 4873 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7qzmb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(29423166-c350-44f8-97ff-adb8b2b99165): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 08:57:48 crc kubenswrapper[4873]: E1201 08:57:48.674295 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="29423166-c350-44f8-97ff-adb8b2b99165" Dec 01 08:57:49 crc kubenswrapper[4873]: E1201 08:57:49.106517 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="29423166-c350-44f8-97ff-adb8b2b99165" Dec 01 08:57:49 crc kubenswrapper[4873]: E1201 08:57:49.107822 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-0" podUID="6d27dec2-61e5-44ae-81df-8fee227ce3d8" Dec 01 08:57:49 crc kubenswrapper[4873]: I1201 08:57:49.110448 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 01 08:57:49 crc kubenswrapper[4873]: W1201 08:57:49.644722 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e6c09a7_97f9_4048_b5a9_b3df685eacbc.slice/crio-96219d362ff82f535ef928ae7c1afc9dd02c140e090a0271d459334975d731df WatchSource:0}: Error finding container 96219d362ff82f535ef928ae7c1afc9dd02c140e090a0271d459334975d731df: Status 404 returned error can't find the container with id 96219d362ff82f535ef928ae7c1afc9dd02c140e090a0271d459334975d731df Dec 01 08:57:49 crc kubenswrapper[4873]: E1201 08:57:49.660684 4873 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 01 08:57:49 crc kubenswrapper[4873]: E1201 08:57:49.660937 4873 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rl9bf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-hssdh_openstack(2a6dc36f-94eb-45a1-b0cd-4f351cd01b1e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 08:57:49 crc kubenswrapper[4873]: E1201 08:57:49.662465 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-hssdh" podUID="2a6dc36f-94eb-45a1-b0cd-4f351cd01b1e" Dec 01 08:57:49 crc kubenswrapper[4873]: E1201 08:57:49.676306 4873 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 01 08:57:49 crc kubenswrapper[4873]: E1201 08:57:49.676582 4873 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-449gp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-jxpr9_openstack(4e867ef7-6308-48b5-b070-fa0ba4513e7f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 08:57:49 crc kubenswrapper[4873]: E1201 08:57:49.677797 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-jxpr9" podUID="4e867ef7-6308-48b5-b070-fa0ba4513e7f" Dec 01 08:57:49 crc kubenswrapper[4873]: E1201 08:57:49.691619 4873 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 01 08:57:49 crc kubenswrapper[4873]: E1201 08:57:49.691864 4873 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-smnfg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-zng4c_openstack(d5588863-2c50-4860-ac45-4d226f59f4ab): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 08:57:49 crc kubenswrapper[4873]: E1201 08:57:49.693266 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-zng4c" podUID="d5588863-2c50-4860-ac45-4d226f59f4ab" Dec 01 08:57:50 crc kubenswrapper[4873]: I1201 08:57:50.186666 4873 generic.go:334] "Generic (PLEG): container finished" podID="34cbfd49-8920-4490-9f8a-7a6f2542d1c6" containerID="1a8c12fd81b743f8d8e4e6b676148b2a4c7f62c53371bb478d28fc5af2831d19" exitCode=0 Dec 01 08:57:50 crc kubenswrapper[4873]: I1201 08:57:50.187432 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-gpkrq" event={"ID":"34cbfd49-8920-4490-9f8a-7a6f2542d1c6","Type":"ContainerDied","Data":"1a8c12fd81b743f8d8e4e6b676148b2a4c7f62c53371bb478d28fc5af2831d19"} Dec 01 08:57:50 crc kubenswrapper[4873]: I1201 08:57:50.205929 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"2e6c09a7-97f9-4048-b5a9-b3df685eacbc","Type":"ContainerStarted","Data":"96219d362ff82f535ef928ae7c1afc9dd02c140e090a0271d459334975d731df"} Dec 01 08:57:50 crc kubenswrapper[4873]: I1201 08:57:50.334982 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-w8v6j"] Dec 01 08:57:50 crc kubenswrapper[4873]: W1201 08:57:50.764571 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod250cbd69_c69a_475c_8973_bc98caf4b264.slice/crio-5e6bcd848d1ed95e563b05b174a6fe93b1781b8f1f6ca25ef0c42e788fa818d4 WatchSource:0}: Error finding container 5e6bcd848d1ed95e563b05b174a6fe93b1781b8f1f6ca25ef0c42e788fa818d4: Status 404 returned error can't find the container with id 5e6bcd848d1ed95e563b05b174a6fe93b1781b8f1f6ca25ef0c42e788fa818d4 Dec 01 08:57:50 crc kubenswrapper[4873]: I1201 08:57:50.792105 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 01 08:57:50 crc kubenswrapper[4873]: I1201 08:57:50.792146 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 01 08:57:50 crc kubenswrapper[4873]: I1201 08:57:50.864918 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 08:57:51 crc kubenswrapper[4873]: W1201 08:57:51.068773 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod21c27fff_80e6_4a74_aad2_5e772b811535.slice/crio-0ae59292458f65c725094ea12c0522e0f25d848464cdd2b157debd62ca4638c8 WatchSource:0}: Error finding container 0ae59292458f65c725094ea12c0522e0f25d848464cdd2b157debd62ca4638c8: Status 404 returned error can't find the container with id 0ae59292458f65c725094ea12c0522e0f25d848464cdd2b157debd62ca4638c8 Dec 01 08:57:51 crc kubenswrapper[4873]: I1201 08:57:51.073810 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 01 08:57:51 crc kubenswrapper[4873]: I1201 08:57:51.236752 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"21c27fff-80e6-4a74-aad2-5e772b811535","Type":"ContainerStarted","Data":"0ae59292458f65c725094ea12c0522e0f25d848464cdd2b157debd62ca4638c8"} Dec 01 08:57:51 crc kubenswrapper[4873]: I1201 08:57:51.250263 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"714ffca1-ee33-4f92-a55e-29e2edf84eff","Type":"ContainerStarted","Data":"dea3d06afdcd538d6458494b19acae8e8c11723a6e337edca6a5ffcbcc6012d3"} Dec 01 08:57:51 crc kubenswrapper[4873]: I1201 08:57:51.259278 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c0ae1a05-ad24-4459-9a9c-1f4993efea08","Type":"ContainerStarted","Data":"65a28273c690248b4e441b31cae5ccb257f159f09dd01ecc4b8ac214beb89190"} Dec 01 08:57:51 crc kubenswrapper[4873]: I1201 08:57:51.266641 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"250cbd69-c69a-475c-8973-bc98caf4b264","Type":"ContainerStarted","Data":"5e6bcd848d1ed95e563b05b174a6fe93b1781b8f1f6ca25ef0c42e788fa818d4"} Dec 01 08:57:51 crc kubenswrapper[4873]: I1201 08:57:51.284914 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-w8v6j" event={"ID":"b79364c5-00b2-4ab8-b2f0-aff17ed902b4","Type":"ContainerStarted","Data":"c53a138a562cbf7e13e1c9317e83de6f2d0e8f3a28b5cd5016821fba0e92e159"} Dec 01 08:57:51 crc kubenswrapper[4873]: E1201 08:57:51.320817 4873 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Dec 01 08:57:51 crc kubenswrapper[4873]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/34cbfd49-8920-4490-9f8a-7a6f2542d1c6/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Dec 01 08:57:51 crc kubenswrapper[4873]: > podSandboxID="33f8d169979df01a49e2cc720645b8540426da2ba6056033c1226b45556d345c" Dec 01 08:57:51 crc kubenswrapper[4873]: E1201 08:57:51.321062 4873 kuberuntime_manager.go:1274] "Unhandled Error" err=< Dec 01 08:57:51 crc kubenswrapper[4873]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5tzzd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-gpkrq_openstack(34cbfd49-8920-4490-9f8a-7a6f2542d1c6): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/34cbfd49-8920-4490-9f8a-7a6f2542d1c6/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Dec 01 08:57:51 crc kubenswrapper[4873]: > logger="UnhandledError" Dec 01 08:57:51 crc kubenswrapper[4873]: E1201 08:57:51.322651 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/34cbfd49-8920-4490-9f8a-7a6f2542d1c6/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-666b6646f7-gpkrq" podUID="34cbfd49-8920-4490-9f8a-7a6f2542d1c6" Dec 01 08:57:51 crc kubenswrapper[4873]: I1201 08:57:51.430525 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 01 08:57:51 crc kubenswrapper[4873]: W1201 08:57:51.447949 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddbe0e07d_50a3_47c2_831d_69f24c0e838a.slice/crio-5fe4bb891029cd4e8e9f2a71aaff74e9c51d7cd0d044c9a1b17fd6ca1b50881f WatchSource:0}: Error finding container 5fe4bb891029cd4e8e9f2a71aaff74e9c51d7cd0d044c9a1b17fd6ca1b50881f: Status 404 returned error can't find the container with id 5fe4bb891029cd4e8e9f2a71aaff74e9c51d7cd0d044c9a1b17fd6ca1b50881f Dec 01 08:57:51 crc kubenswrapper[4873]: I1201 08:57:51.508238 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-jxpr9" Dec 01 08:57:51 crc kubenswrapper[4873]: I1201 08:57:51.509749 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-xd66n"] Dec 01 08:57:51 crc kubenswrapper[4873]: I1201 08:57:51.515272 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-hssdh" Dec 01 08:57:51 crc kubenswrapper[4873]: I1201 08:57:51.530328 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e867ef7-6308-48b5-b070-fa0ba4513e7f-config\") pod \"4e867ef7-6308-48b5-b070-fa0ba4513e7f\" (UID: \"4e867ef7-6308-48b5-b070-fa0ba4513e7f\") " Dec 01 08:57:51 crc kubenswrapper[4873]: I1201 08:57:51.530404 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2a6dc36f-94eb-45a1-b0cd-4f351cd01b1e-dns-svc\") pod \"2a6dc36f-94eb-45a1-b0cd-4f351cd01b1e\" (UID: \"2a6dc36f-94eb-45a1-b0cd-4f351cd01b1e\") " Dec 01 08:57:51 crc kubenswrapper[4873]: I1201 08:57:51.530429 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a6dc36f-94eb-45a1-b0cd-4f351cd01b1e-config\") pod \"2a6dc36f-94eb-45a1-b0cd-4f351cd01b1e\" (UID: \"2a6dc36f-94eb-45a1-b0cd-4f351cd01b1e\") " Dec 01 08:57:51 crc kubenswrapper[4873]: I1201 08:57:51.530451 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-449gp\" (UniqueName: \"kubernetes.io/projected/4e867ef7-6308-48b5-b070-fa0ba4513e7f-kube-api-access-449gp\") pod \"4e867ef7-6308-48b5-b070-fa0ba4513e7f\" (UID: \"4e867ef7-6308-48b5-b070-fa0ba4513e7f\") " Dec 01 08:57:51 crc kubenswrapper[4873]: I1201 08:57:51.530472 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rl9bf\" (UniqueName: \"kubernetes.io/projected/2a6dc36f-94eb-45a1-b0cd-4f351cd01b1e-kube-api-access-rl9bf\") pod \"2a6dc36f-94eb-45a1-b0cd-4f351cd01b1e\" (UID: \"2a6dc36f-94eb-45a1-b0cd-4f351cd01b1e\") " Dec 01 08:57:51 crc kubenswrapper[4873]: I1201 08:57:51.534866 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a6dc36f-94eb-45a1-b0cd-4f351cd01b1e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2a6dc36f-94eb-45a1-b0cd-4f351cd01b1e" (UID: "2a6dc36f-94eb-45a1-b0cd-4f351cd01b1e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:57:51 crc kubenswrapper[4873]: I1201 08:57:51.535490 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a6dc36f-94eb-45a1-b0cd-4f351cd01b1e-config" (OuterVolumeSpecName: "config") pod "2a6dc36f-94eb-45a1-b0cd-4f351cd01b1e" (UID: "2a6dc36f-94eb-45a1-b0cd-4f351cd01b1e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:57:51 crc kubenswrapper[4873]: I1201 08:57:51.541245 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a6dc36f-94eb-45a1-b0cd-4f351cd01b1e-kube-api-access-rl9bf" (OuterVolumeSpecName: "kube-api-access-rl9bf") pod "2a6dc36f-94eb-45a1-b0cd-4f351cd01b1e" (UID: "2a6dc36f-94eb-45a1-b0cd-4f351cd01b1e"). InnerVolumeSpecName "kube-api-access-rl9bf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:57:51 crc kubenswrapper[4873]: I1201 08:57:51.541285 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e867ef7-6308-48b5-b070-fa0ba4513e7f-kube-api-access-449gp" (OuterVolumeSpecName: "kube-api-access-449gp") pod "4e867ef7-6308-48b5-b070-fa0ba4513e7f" (UID: "4e867ef7-6308-48b5-b070-fa0ba4513e7f"). InnerVolumeSpecName "kube-api-access-449gp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:57:51 crc kubenswrapper[4873]: I1201 08:57:51.551123 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e867ef7-6308-48b5-b070-fa0ba4513e7f-config" (OuterVolumeSpecName: "config") pod "4e867ef7-6308-48b5-b070-fa0ba4513e7f" (UID: "4e867ef7-6308-48b5-b070-fa0ba4513e7f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:57:51 crc kubenswrapper[4873]: I1201 08:57:51.632311 4873 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e867ef7-6308-48b5-b070-fa0ba4513e7f-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:57:51 crc kubenswrapper[4873]: I1201 08:57:51.632352 4873 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2a6dc36f-94eb-45a1-b0cd-4f351cd01b1e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 08:57:51 crc kubenswrapper[4873]: I1201 08:57:51.632362 4873 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a6dc36f-94eb-45a1-b0cd-4f351cd01b1e-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:57:51 crc kubenswrapper[4873]: I1201 08:57:51.632376 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-449gp\" (UniqueName: \"kubernetes.io/projected/4e867ef7-6308-48b5-b070-fa0ba4513e7f-kube-api-access-449gp\") on node \"crc\" DevicePath \"\"" Dec 01 08:57:51 crc kubenswrapper[4873]: I1201 08:57:51.632388 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rl9bf\" (UniqueName: \"kubernetes.io/projected/2a6dc36f-94eb-45a1-b0cd-4f351cd01b1e-kube-api-access-rl9bf\") on node \"crc\" DevicePath \"\"" Dec 01 08:57:52 crc kubenswrapper[4873]: I1201 08:57:52.300623 4873 generic.go:334] "Generic (PLEG): container finished" podID="d5588863-2c50-4860-ac45-4d226f59f4ab" containerID="16ad9f3f5dbafb64151afe4ad7fd52500cb0482d633e2933de657e75ecc5105f" exitCode=0 Dec 01 08:57:52 crc kubenswrapper[4873]: I1201 08:57:52.300883 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-zng4c" event={"ID":"d5588863-2c50-4860-ac45-4d226f59f4ab","Type":"ContainerDied","Data":"16ad9f3f5dbafb64151afe4ad7fd52500cb0482d633e2933de657e75ecc5105f"} Dec 01 08:57:52 crc kubenswrapper[4873]: I1201 08:57:52.311820 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-xd66n" event={"ID":"243fec55-bb19-4807-8b15-69584a922cf6","Type":"ContainerStarted","Data":"5a327e35d90d575e09bc59f5af16101f3408e7fdb935230329f0f6a416e37a94"} Dec 01 08:57:52 crc kubenswrapper[4873]: I1201 08:57:52.314734 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-jxpr9" event={"ID":"4e867ef7-6308-48b5-b070-fa0ba4513e7f","Type":"ContainerDied","Data":"745447136c68f7001636207498c3efaecaaee554488c7b7294ff861975dcc754"} Dec 01 08:57:52 crc kubenswrapper[4873]: I1201 08:57:52.314833 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-jxpr9" Dec 01 08:57:52 crc kubenswrapper[4873]: I1201 08:57:52.317400 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"dbe0e07d-50a3-47c2-831d-69f24c0e838a","Type":"ContainerStarted","Data":"5fe4bb891029cd4e8e9f2a71aaff74e9c51d7cd0d044c9a1b17fd6ca1b50881f"} Dec 01 08:57:52 crc kubenswrapper[4873]: I1201 08:57:52.318468 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-hssdh" Dec 01 08:57:52 crc kubenswrapper[4873]: I1201 08:57:52.319498 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-hssdh" event={"ID":"2a6dc36f-94eb-45a1-b0cd-4f351cd01b1e","Type":"ContainerDied","Data":"d7b3e03ad2a06029fad081c99cdcddee4d95212b11212a4044e95a42f693ade3"} Dec 01 08:57:52 crc kubenswrapper[4873]: I1201 08:57:52.456718 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-jxpr9"] Dec 01 08:57:52 crc kubenswrapper[4873]: I1201 08:57:52.465204 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-jxpr9"] Dec 01 08:57:52 crc kubenswrapper[4873]: I1201 08:57:52.489674 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-hssdh"] Dec 01 08:57:52 crc kubenswrapper[4873]: I1201 08:57:52.496873 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-hssdh"] Dec 01 08:57:54 crc kubenswrapper[4873]: I1201 08:57:54.445351 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a6dc36f-94eb-45a1-b0cd-4f351cd01b1e" path="/var/lib/kubelet/pods/2a6dc36f-94eb-45a1-b0cd-4f351cd01b1e/volumes" Dec 01 08:57:54 crc kubenswrapper[4873]: I1201 08:57:54.446335 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e867ef7-6308-48b5-b070-fa0ba4513e7f" path="/var/lib/kubelet/pods/4e867ef7-6308-48b5-b070-fa0ba4513e7f/volumes" Dec 01 08:57:59 crc kubenswrapper[4873]: I1201 08:57:59.412614 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-gpkrq" event={"ID":"34cbfd49-8920-4490-9f8a-7a6f2542d1c6","Type":"ContainerStarted","Data":"5a1f8f98f1c2680b702cd2ff1d7aedc07c140dd8eda4a5cc515f58974915bc70"} Dec 01 08:57:59 crc kubenswrapper[4873]: I1201 08:57:59.414876 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-gpkrq" Dec 01 08:57:59 crc kubenswrapper[4873]: I1201 08:57:59.421862 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-zng4c" event={"ID":"d5588863-2c50-4860-ac45-4d226f59f4ab","Type":"ContainerStarted","Data":"99d10df78458a4b918d3e158949e5fd8107d001f23798facc6d8904c218ffc35"} Dec 01 08:57:59 crc kubenswrapper[4873]: I1201 08:57:59.423126 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-zng4c" Dec 01 08:57:59 crc kubenswrapper[4873]: I1201 08:57:59.432071 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"2e6c09a7-97f9-4048-b5a9-b3df685eacbc","Type":"ContainerStarted","Data":"c2b7281207419f96164147d82809ea20b83c1ef67f3f51488929ea2fdff3e9fa"} Dec 01 08:57:59 crc kubenswrapper[4873]: I1201 08:57:59.432436 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 01 08:57:59 crc kubenswrapper[4873]: I1201 08:57:59.441310 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"dbe0e07d-50a3-47c2-831d-69f24c0e838a","Type":"ContainerStarted","Data":"d6804045a5b52fd8486a10e8d8ec1fbea3765d18ebb6ce9dfc038a4d48913a7f"} Dec 01 08:57:59 crc kubenswrapper[4873]: I1201 08:57:59.451948 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-gpkrq" podStartSLOduration=11.441010472 podStartE2EDuration="37.451916433s" podCreationTimestamp="2025-12-01 08:57:22 +0000 UTC" firstStartedPulling="2025-12-01 08:57:23.808905941 +0000 UTC m=+1019.711014470" lastFinishedPulling="2025-12-01 08:57:49.819811892 +0000 UTC m=+1045.721920431" observedRunningTime="2025-12-01 08:57:59.444767294 +0000 UTC m=+1055.346875843" watchObservedRunningTime="2025-12-01 08:57:59.451916433 +0000 UTC m=+1055.354024972" Dec 01 08:57:59 crc kubenswrapper[4873]: I1201 08:57:59.502524 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-zng4c" podStartSLOduration=-9223372000.352278 podStartE2EDuration="36.502498376s" podCreationTimestamp="2025-12-01 08:57:23 +0000 UTC" firstStartedPulling="2025-12-01 08:57:23.857923387 +0000 UTC m=+1019.760031916" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:57:59.496484664 +0000 UTC m=+1055.398593193" watchObservedRunningTime="2025-12-01 08:57:59.502498376 +0000 UTC m=+1055.404606915" Dec 01 08:57:59 crc kubenswrapper[4873]: I1201 08:57:59.532168 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=24.079203758 podStartE2EDuration="32.532141245s" podCreationTimestamp="2025-12-01 08:57:27 +0000 UTC" firstStartedPulling="2025-12-01 08:57:49.693254218 +0000 UTC m=+1045.595362767" lastFinishedPulling="2025-12-01 08:57:58.146191715 +0000 UTC m=+1054.048300254" observedRunningTime="2025-12-01 08:57:59.531772836 +0000 UTC m=+1055.433881375" watchObservedRunningTime="2025-12-01 08:57:59.532141245 +0000 UTC m=+1055.434249784" Dec 01 08:58:01 crc kubenswrapper[4873]: I1201 08:58:01.482788 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c0ae1a05-ad24-4459-9a9c-1f4993efea08","Type":"ContainerStarted","Data":"71ffe686e209a3865545af99dca115be6ab26bd5686728be1ce81bf5da6a4107"} Dec 01 08:58:01 crc kubenswrapper[4873]: I1201 08:58:01.483219 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 01 08:58:01 crc kubenswrapper[4873]: I1201 08:58:01.486910 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"250cbd69-c69a-475c-8973-bc98caf4b264","Type":"ContainerStarted","Data":"5eb0007bd3fb694a8a3489c551954b5beb7f2b900b7bef9c3ec657938d95e850"} Dec 01 08:58:01 crc kubenswrapper[4873]: I1201 08:58:01.489531 4873 generic.go:334] "Generic (PLEG): container finished" podID="243fec55-bb19-4807-8b15-69584a922cf6" containerID="617a605e344e518c0034cbf1955e4cef3ed31ba251ada8837dc44ac118e3a263" exitCode=0 Dec 01 08:58:01 crc kubenswrapper[4873]: I1201 08:58:01.489595 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-xd66n" event={"ID":"243fec55-bb19-4807-8b15-69584a922cf6","Type":"ContainerDied","Data":"617a605e344e518c0034cbf1955e4cef3ed31ba251ada8837dc44ac118e3a263"} Dec 01 08:58:01 crc kubenswrapper[4873]: I1201 08:58:01.496703 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"21c27fff-80e6-4a74-aad2-5e772b811535","Type":"ContainerStarted","Data":"5dc759b21b6c379eef52e814787eaffd0d3f4c26158d46920ae069758e6473e0"} Dec 01 08:58:01 crc kubenswrapper[4873]: I1201 08:58:01.499904 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-w8v6j" event={"ID":"b79364c5-00b2-4ab8-b2f0-aff17ed902b4","Type":"ContainerStarted","Data":"b321b24ba79fe8db2c85b86b82516fe0cc565f05376aefc69c414904893f12f7"} Dec 01 08:58:01 crc kubenswrapper[4873]: I1201 08:58:01.501068 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-w8v6j" Dec 01 08:58:01 crc kubenswrapper[4873]: I1201 08:58:01.507254 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"714ffca1-ee33-4f92-a55e-29e2edf84eff","Type":"ContainerStarted","Data":"b75e73cd37b2f144991507973eae59641201472a490d1c34e73f6530b942591f"} Dec 01 08:58:01 crc kubenswrapper[4873]: I1201 08:58:01.549182 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=25.584287301 podStartE2EDuration="33.549158902s" podCreationTimestamp="2025-12-01 08:57:28 +0000 UTC" firstStartedPulling="2025-12-01 08:57:51.062031711 +0000 UTC m=+1046.964140250" lastFinishedPulling="2025-12-01 08:57:59.026903312 +0000 UTC m=+1054.929011851" observedRunningTime="2025-12-01 08:58:01.544073452 +0000 UTC m=+1057.446181981" watchObservedRunningTime="2025-12-01 08:58:01.549158902 +0000 UTC m=+1057.451267431" Dec 01 08:58:01 crc kubenswrapper[4873]: I1201 08:58:01.654877 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-w8v6j" podStartSLOduration=21.736406041 podStartE2EDuration="29.654851564s" podCreationTimestamp="2025-12-01 08:57:32 +0000 UTC" firstStartedPulling="2025-12-01 08:57:50.624486747 +0000 UTC m=+1046.526595286" lastFinishedPulling="2025-12-01 08:57:58.54293227 +0000 UTC m=+1054.445040809" observedRunningTime="2025-12-01 08:58:01.649470157 +0000 UTC m=+1057.551578696" watchObservedRunningTime="2025-12-01 08:58:01.654851564 +0000 UTC m=+1057.556960103" Dec 01 08:58:03 crc kubenswrapper[4873]: I1201 08:58:03.557595 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"dbe0e07d-50a3-47c2-831d-69f24c0e838a","Type":"ContainerStarted","Data":"a91072fa7cb95af98c3f4c9d7599f062e6add266aabcc89cdcd39f775e7fbbe3"} Dec 01 08:58:03 crc kubenswrapper[4873]: I1201 08:58:03.560911 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-xd66n" event={"ID":"243fec55-bb19-4807-8b15-69584a922cf6","Type":"ContainerStarted","Data":"a468286085cbd5853586ce0e79a8f27f238dbf7d57c6e797c214b16da9c5ebfa"} Dec 01 08:58:03 crc kubenswrapper[4873]: I1201 08:58:03.563758 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"21c27fff-80e6-4a74-aad2-5e772b811535","Type":"ContainerStarted","Data":"4869eb38acd6be8b0bc3773e45c0a41075627740f7a3c16d218479b1a5843853"} Dec 01 08:58:03 crc kubenswrapper[4873]: I1201 08:58:03.612746 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=17.002475776 podStartE2EDuration="28.612717629s" podCreationTimestamp="2025-12-01 08:57:35 +0000 UTC" firstStartedPulling="2025-12-01 08:57:51.451078494 +0000 UTC m=+1047.353187033" lastFinishedPulling="2025-12-01 08:58:03.061320347 +0000 UTC m=+1058.963428886" observedRunningTime="2025-12-01 08:58:03.582640549 +0000 UTC m=+1059.484749088" watchObservedRunningTime="2025-12-01 08:58:03.612717629 +0000 UTC m=+1059.514826168" Dec 01 08:58:03 crc kubenswrapper[4873]: I1201 08:58:03.615496 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=19.621310119 podStartE2EDuration="31.615487494s" podCreationTimestamp="2025-12-01 08:57:32 +0000 UTC" firstStartedPulling="2025-12-01 08:57:51.072111059 +0000 UTC m=+1046.974219598" lastFinishedPulling="2025-12-01 08:58:03.066288444 +0000 UTC m=+1058.968396973" observedRunningTime="2025-12-01 08:58:03.609632696 +0000 UTC m=+1059.511741235" watchObservedRunningTime="2025-12-01 08:58:03.615487494 +0000 UTC m=+1059.517596033" Dec 01 08:58:03 crc kubenswrapper[4873]: I1201 08:58:03.860773 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 01 08:58:03 crc kubenswrapper[4873]: I1201 08:58:03.899491 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 01 08:58:04 crc kubenswrapper[4873]: I1201 08:58:04.415669 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 01 08:58:04 crc kubenswrapper[4873]: I1201 08:58:04.415725 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 01 08:58:04 crc kubenswrapper[4873]: I1201 08:58:04.480886 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 01 08:58:04 crc kubenswrapper[4873]: I1201 08:58:04.576114 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-xd66n" event={"ID":"243fec55-bb19-4807-8b15-69584a922cf6","Type":"ContainerStarted","Data":"70ec324da183c84ee573703e777bf44563ea4293ff63eab5aac85d8396dada4d"} Dec 01 08:58:04 crc kubenswrapper[4873]: I1201 08:58:04.576185 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-xd66n" Dec 01 08:58:04 crc kubenswrapper[4873]: I1201 08:58:04.577594 4873 generic.go:334] "Generic (PLEG): container finished" podID="714ffca1-ee33-4f92-a55e-29e2edf84eff" containerID="b75e73cd37b2f144991507973eae59641201472a490d1c34e73f6530b942591f" exitCode=0 Dec 01 08:58:04 crc kubenswrapper[4873]: I1201 08:58:04.577657 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"714ffca1-ee33-4f92-a55e-29e2edf84eff","Type":"ContainerDied","Data":"b75e73cd37b2f144991507973eae59641201472a490d1c34e73f6530b942591f"} Dec 01 08:58:04 crc kubenswrapper[4873]: I1201 08:58:04.578125 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 01 08:58:04 crc kubenswrapper[4873]: I1201 08:58:04.842091 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 01 08:58:04 crc kubenswrapper[4873]: I1201 08:58:04.877693 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-xd66n" podStartSLOduration=25.08568177 podStartE2EDuration="31.877653174s" podCreationTimestamp="2025-12-01 08:57:33 +0000 UTC" firstStartedPulling="2025-12-01 08:57:51.515164296 +0000 UTC m=+1047.417272835" lastFinishedPulling="2025-12-01 08:57:58.3071357 +0000 UTC m=+1054.209244239" observedRunningTime="2025-12-01 08:58:04.848690261 +0000 UTC m=+1060.750798800" watchObservedRunningTime="2025-12-01 08:58:04.877653174 +0000 UTC m=+1060.779761713" Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.406919 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-gpkrq"] Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.407591 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-gpkrq" podUID="34cbfd49-8920-4490-9f8a-7a6f2542d1c6" containerName="dnsmasq-dns" containerID="cri-o://5a1f8f98f1c2680b702cd2ff1d7aedc07c140dd8eda4a5cc515f58974915bc70" gracePeriod=10 Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.410273 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-666b6646f7-gpkrq" Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.550856 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-7xdqf"] Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.552323 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-x2trf"] Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.553373 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-x2trf" Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.556604 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-7xdqf" Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.563137 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.563469 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.633836 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35124c22-9257-440f-8a97-9e029187bd69-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-x2trf\" (UID: \"35124c22-9257-440f-8a97-9e029187bd69\") " pod="openstack/dnsmasq-dns-7f896c8c65-x2trf" Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.633917 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35124c22-9257-440f-8a97-9e029187bd69-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-x2trf\" (UID: \"35124c22-9257-440f-8a97-9e029187bd69\") " pod="openstack/dnsmasq-dns-7f896c8c65-x2trf" Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.633942 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/93fb09e4-51b3-44f7-afab-c49c374c7f56-ovs-rundir\") pod \"ovn-controller-metrics-7xdqf\" (UID: \"93fb09e4-51b3-44f7-afab-c49c374c7f56\") " pod="openstack/ovn-controller-metrics-7xdqf" Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.633995 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fh97\" (UniqueName: \"kubernetes.io/projected/93fb09e4-51b3-44f7-afab-c49c374c7f56-kube-api-access-6fh97\") pod \"ovn-controller-metrics-7xdqf\" (UID: \"93fb09e4-51b3-44f7-afab-c49c374c7f56\") " pod="openstack/ovn-controller-metrics-7xdqf" Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.634025 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/93fb09e4-51b3-44f7-afab-c49c374c7f56-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-7xdqf\" (UID: \"93fb09e4-51b3-44f7-afab-c49c374c7f56\") " pod="openstack/ovn-controller-metrics-7xdqf" Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.634062 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93fb09e4-51b3-44f7-afab-c49c374c7f56-config\") pod \"ovn-controller-metrics-7xdqf\" (UID: \"93fb09e4-51b3-44f7-afab-c49c374c7f56\") " pod="openstack/ovn-controller-metrics-7xdqf" Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.634099 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93fb09e4-51b3-44f7-afab-c49c374c7f56-combined-ca-bundle\") pod \"ovn-controller-metrics-7xdqf\" (UID: \"93fb09e4-51b3-44f7-afab-c49c374c7f56\") " pod="openstack/ovn-controller-metrics-7xdqf" Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.634170 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/93fb09e4-51b3-44f7-afab-c49c374c7f56-ovn-rundir\") pod \"ovn-controller-metrics-7xdqf\" (UID: \"93fb09e4-51b3-44f7-afab-c49c374c7f56\") " pod="openstack/ovn-controller-metrics-7xdqf" Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.634215 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35124c22-9257-440f-8a97-9e029187bd69-config\") pod \"dnsmasq-dns-7f896c8c65-x2trf\" (UID: \"35124c22-9257-440f-8a97-9e029187bd69\") " pod="openstack/dnsmasq-dns-7f896c8c65-x2trf" Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.634243 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4449\" (UniqueName: \"kubernetes.io/projected/35124c22-9257-440f-8a97-9e029187bd69-kube-api-access-f4449\") pod \"dnsmasq-dns-7f896c8c65-x2trf\" (UID: \"35124c22-9257-440f-8a97-9e029187bd69\") " pod="openstack/dnsmasq-dns-7f896c8c65-x2trf" Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.672129 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-x2trf"] Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.743915 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-7xdqf"] Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.744864 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35124c22-9257-440f-8a97-9e029187bd69-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-x2trf\" (UID: \"35124c22-9257-440f-8a97-9e029187bd69\") " pod="openstack/dnsmasq-dns-7f896c8c65-x2trf" Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.744895 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/93fb09e4-51b3-44f7-afab-c49c374c7f56-ovs-rundir\") pod \"ovn-controller-metrics-7xdqf\" (UID: \"93fb09e4-51b3-44f7-afab-c49c374c7f56\") " pod="openstack/ovn-controller-metrics-7xdqf" Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.744946 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fh97\" (UniqueName: \"kubernetes.io/projected/93fb09e4-51b3-44f7-afab-c49c374c7f56-kube-api-access-6fh97\") pod \"ovn-controller-metrics-7xdqf\" (UID: \"93fb09e4-51b3-44f7-afab-c49c374c7f56\") " pod="openstack/ovn-controller-metrics-7xdqf" Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.744969 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/93fb09e4-51b3-44f7-afab-c49c374c7f56-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-7xdqf\" (UID: \"93fb09e4-51b3-44f7-afab-c49c374c7f56\") " pod="openstack/ovn-controller-metrics-7xdqf" Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.744990 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93fb09e4-51b3-44f7-afab-c49c374c7f56-config\") pod \"ovn-controller-metrics-7xdqf\" (UID: \"93fb09e4-51b3-44f7-afab-c49c374c7f56\") " pod="openstack/ovn-controller-metrics-7xdqf" Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.745022 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93fb09e4-51b3-44f7-afab-c49c374c7f56-combined-ca-bundle\") pod \"ovn-controller-metrics-7xdqf\" (UID: \"93fb09e4-51b3-44f7-afab-c49c374c7f56\") " pod="openstack/ovn-controller-metrics-7xdqf" Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.745076 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/93fb09e4-51b3-44f7-afab-c49c374c7f56-ovn-rundir\") pod \"ovn-controller-metrics-7xdqf\" (UID: \"93fb09e4-51b3-44f7-afab-c49c374c7f56\") " pod="openstack/ovn-controller-metrics-7xdqf" Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.745105 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35124c22-9257-440f-8a97-9e029187bd69-config\") pod \"dnsmasq-dns-7f896c8c65-x2trf\" (UID: \"35124c22-9257-440f-8a97-9e029187bd69\") " pod="openstack/dnsmasq-dns-7f896c8c65-x2trf" Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.745128 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4449\" (UniqueName: \"kubernetes.io/projected/35124c22-9257-440f-8a97-9e029187bd69-kube-api-access-f4449\") pod \"dnsmasq-dns-7f896c8c65-x2trf\" (UID: \"35124c22-9257-440f-8a97-9e029187bd69\") " pod="openstack/dnsmasq-dns-7f896c8c65-x2trf" Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.745152 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35124c22-9257-440f-8a97-9e029187bd69-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-x2trf\" (UID: \"35124c22-9257-440f-8a97-9e029187bd69\") " pod="openstack/dnsmasq-dns-7f896c8c65-x2trf" Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.746213 4873 generic.go:334] "Generic (PLEG): container finished" podID="250cbd69-c69a-475c-8973-bc98caf4b264" containerID="5eb0007bd3fb694a8a3489c551954b5beb7f2b900b7bef9c3ec657938d95e850" exitCode=0 Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.746265 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35124c22-9257-440f-8a97-9e029187bd69-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-x2trf\" (UID: \"35124c22-9257-440f-8a97-9e029187bd69\") " pod="openstack/dnsmasq-dns-7f896c8c65-x2trf" Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.746297 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"250cbd69-c69a-475c-8973-bc98caf4b264","Type":"ContainerDied","Data":"5eb0007bd3fb694a8a3489c551954b5beb7f2b900b7bef9c3ec657938d95e850"} Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.750015 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/93fb09e4-51b3-44f7-afab-c49c374c7f56-ovn-rundir\") pod \"ovn-controller-metrics-7xdqf\" (UID: \"93fb09e4-51b3-44f7-afab-c49c374c7f56\") " pod="openstack/ovn-controller-metrics-7xdqf" Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.752548 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/93fb09e4-51b3-44f7-afab-c49c374c7f56-ovs-rundir\") pod \"ovn-controller-metrics-7xdqf\" (UID: \"93fb09e4-51b3-44f7-afab-c49c374c7f56\") " pod="openstack/ovn-controller-metrics-7xdqf" Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.753968 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35124c22-9257-440f-8a97-9e029187bd69-config\") pod \"dnsmasq-dns-7f896c8c65-x2trf\" (UID: \"35124c22-9257-440f-8a97-9e029187bd69\") " pod="openstack/dnsmasq-dns-7f896c8c65-x2trf" Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.757642 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35124c22-9257-440f-8a97-9e029187bd69-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-x2trf\" (UID: \"35124c22-9257-440f-8a97-9e029187bd69\") " pod="openstack/dnsmasq-dns-7f896c8c65-x2trf" Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.760266 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/93fb09e4-51b3-44f7-afab-c49c374c7f56-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-7xdqf\" (UID: \"93fb09e4-51b3-44f7-afab-c49c374c7f56\") " pod="openstack/ovn-controller-metrics-7xdqf" Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.762860 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"714ffca1-ee33-4f92-a55e-29e2edf84eff","Type":"ContainerStarted","Data":"f3dd1e3d08309406feda962daa1d3dcb7dec1afa654962e31071dc7f98802093"} Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.766151 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93fb09e4-51b3-44f7-afab-c49c374c7f56-config\") pod \"ovn-controller-metrics-7xdqf\" (UID: \"93fb09e4-51b3-44f7-afab-c49c374c7f56\") " pod="openstack/ovn-controller-metrics-7xdqf" Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.786541 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6d27dec2-61e5-44ae-81df-8fee227ce3d8","Type":"ContainerStarted","Data":"cfec876654e058bc3231e41fa25327a07bbec96f72ff60c123f47775b7ad0e34"} Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.786983 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fh97\" (UniqueName: \"kubernetes.io/projected/93fb09e4-51b3-44f7-afab-c49c374c7f56-kube-api-access-6fh97\") pod \"ovn-controller-metrics-7xdqf\" (UID: \"93fb09e4-51b3-44f7-afab-c49c374c7f56\") " pod="openstack/ovn-controller-metrics-7xdqf" Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.791804 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93fb09e4-51b3-44f7-afab-c49c374c7f56-combined-ca-bundle\") pod \"ovn-controller-metrics-7xdqf\" (UID: \"93fb09e4-51b3-44f7-afab-c49c374c7f56\") " pod="openstack/ovn-controller-metrics-7xdqf" Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.794481 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"29423166-c350-44f8-97ff-adb8b2b99165","Type":"ContainerStarted","Data":"00ccca11cf2b2f7b96f7e096e15b6f797e1625129d5561faba84ede8dfacb4e9"} Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.794703 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-xd66n" Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.796570 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4449\" (UniqueName: \"kubernetes.io/projected/35124c22-9257-440f-8a97-9e029187bd69-kube-api-access-f4449\") pod \"dnsmasq-dns-7f896c8c65-x2trf\" (UID: \"35124c22-9257-440f-8a97-9e029187bd69\") " pod="openstack/dnsmasq-dns-7f896c8c65-x2trf" Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.866821 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.888989 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-7xdqf" Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.911945 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=32.678557322 podStartE2EDuration="40.91191643s" podCreationTimestamp="2025-12-01 08:57:25 +0000 UTC" firstStartedPulling="2025-12-01 08:57:50.62416865 +0000 UTC m=+1046.526277189" lastFinishedPulling="2025-12-01 08:57:58.857527758 +0000 UTC m=+1054.759636297" observedRunningTime="2025-12-01 08:58:05.866813876 +0000 UTC m=+1061.768922415" watchObservedRunningTime="2025-12-01 08:58:05.91191643 +0000 UTC m=+1061.814024969" Dec 01 08:58:05 crc kubenswrapper[4873]: I1201 08:58:05.979503 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-x2trf" Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.026922 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-zng4c"] Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.035559 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-zng4c" podUID="d5588863-2c50-4860-ac45-4d226f59f4ab" containerName="dnsmasq-dns" containerID="cri-o://99d10df78458a4b918d3e158949e5fd8107d001f23798facc6d8904c218ffc35" gracePeriod=10 Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.041208 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-zng4c" Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.072598 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-j7n86"] Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.074486 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-j7n86" Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.076991 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.112115 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-j7n86"] Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.175887 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjpkk\" (UniqueName: \"kubernetes.io/projected/11ec3abe-6052-445d-9f3f-016157fecbf9-kube-api-access-pjpkk\") pod \"dnsmasq-dns-86db49b7ff-j7n86\" (UID: \"11ec3abe-6052-445d-9f3f-016157fecbf9\") " pod="openstack/dnsmasq-dns-86db49b7ff-j7n86" Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.176445 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11ec3abe-6052-445d-9f3f-016157fecbf9-config\") pod \"dnsmasq-dns-86db49b7ff-j7n86\" (UID: \"11ec3abe-6052-445d-9f3f-016157fecbf9\") " pod="openstack/dnsmasq-dns-86db49b7ff-j7n86" Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.176492 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11ec3abe-6052-445d-9f3f-016157fecbf9-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-j7n86\" (UID: \"11ec3abe-6052-445d-9f3f-016157fecbf9\") " pod="openstack/dnsmasq-dns-86db49b7ff-j7n86" Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.176532 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/11ec3abe-6052-445d-9f3f-016157fecbf9-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-j7n86\" (UID: \"11ec3abe-6052-445d-9f3f-016157fecbf9\") " pod="openstack/dnsmasq-dns-86db49b7ff-j7n86" Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.176600 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11ec3abe-6052-445d-9f3f-016157fecbf9-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-j7n86\" (UID: \"11ec3abe-6052-445d-9f3f-016157fecbf9\") " pod="openstack/dnsmasq-dns-86db49b7ff-j7n86" Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.278714 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjpkk\" (UniqueName: \"kubernetes.io/projected/11ec3abe-6052-445d-9f3f-016157fecbf9-kube-api-access-pjpkk\") pod \"dnsmasq-dns-86db49b7ff-j7n86\" (UID: \"11ec3abe-6052-445d-9f3f-016157fecbf9\") " pod="openstack/dnsmasq-dns-86db49b7ff-j7n86" Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.278806 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11ec3abe-6052-445d-9f3f-016157fecbf9-config\") pod \"dnsmasq-dns-86db49b7ff-j7n86\" (UID: \"11ec3abe-6052-445d-9f3f-016157fecbf9\") " pod="openstack/dnsmasq-dns-86db49b7ff-j7n86" Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.278841 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11ec3abe-6052-445d-9f3f-016157fecbf9-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-j7n86\" (UID: \"11ec3abe-6052-445d-9f3f-016157fecbf9\") " pod="openstack/dnsmasq-dns-86db49b7ff-j7n86" Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.278881 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/11ec3abe-6052-445d-9f3f-016157fecbf9-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-j7n86\" (UID: \"11ec3abe-6052-445d-9f3f-016157fecbf9\") " pod="openstack/dnsmasq-dns-86db49b7ff-j7n86" Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.278947 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11ec3abe-6052-445d-9f3f-016157fecbf9-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-j7n86\" (UID: \"11ec3abe-6052-445d-9f3f-016157fecbf9\") " pod="openstack/dnsmasq-dns-86db49b7ff-j7n86" Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.280443 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11ec3abe-6052-445d-9f3f-016157fecbf9-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-j7n86\" (UID: \"11ec3abe-6052-445d-9f3f-016157fecbf9\") " pod="openstack/dnsmasq-dns-86db49b7ff-j7n86" Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.280475 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11ec3abe-6052-445d-9f3f-016157fecbf9-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-j7n86\" (UID: \"11ec3abe-6052-445d-9f3f-016157fecbf9\") " pod="openstack/dnsmasq-dns-86db49b7ff-j7n86" Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.280569 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11ec3abe-6052-445d-9f3f-016157fecbf9-config\") pod \"dnsmasq-dns-86db49b7ff-j7n86\" (UID: \"11ec3abe-6052-445d-9f3f-016157fecbf9\") " pod="openstack/dnsmasq-dns-86db49b7ff-j7n86" Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.281007 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/11ec3abe-6052-445d-9f3f-016157fecbf9-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-j7n86\" (UID: \"11ec3abe-6052-445d-9f3f-016157fecbf9\") " pod="openstack/dnsmasq-dns-86db49b7ff-j7n86" Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.322305 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjpkk\" (UniqueName: \"kubernetes.io/projected/11ec3abe-6052-445d-9f3f-016157fecbf9-kube-api-access-pjpkk\") pod \"dnsmasq-dns-86db49b7ff-j7n86\" (UID: \"11ec3abe-6052-445d-9f3f-016157fecbf9\") " pod="openstack/dnsmasq-dns-86db49b7ff-j7n86" Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.456740 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-j7n86" Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.487225 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-7xdqf"] Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.543993 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.546522 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.795351 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-wqh9w" Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.795604 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.795723 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.796266 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.802301 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.809925 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5cf24b4-dd05-45a4-bf8f-5895c30ada04-config\") pod \"ovn-northd-0\" (UID: \"a5cf24b4-dd05-45a4-bf8f-5895c30ada04\") " pod="openstack/ovn-northd-0" Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.810009 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5cf24b4-dd05-45a4-bf8f-5895c30ada04-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"a5cf24b4-dd05-45a4-bf8f-5895c30ada04\") " pod="openstack/ovn-northd-0" Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.810097 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5cf24b4-dd05-45a4-bf8f-5895c30ada04-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"a5cf24b4-dd05-45a4-bf8f-5895c30ada04\") " pod="openstack/ovn-northd-0" Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.810238 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cf24b4-dd05-45a4-bf8f-5895c30ada04-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"a5cf24b4-dd05-45a4-bf8f-5895c30ada04\") " pod="openstack/ovn-northd-0" Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.810258 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fv2jn\" (UniqueName: \"kubernetes.io/projected/a5cf24b4-dd05-45a4-bf8f-5895c30ada04-kube-api-access-fv2jn\") pod \"ovn-northd-0\" (UID: \"a5cf24b4-dd05-45a4-bf8f-5895c30ada04\") " pod="openstack/ovn-northd-0" Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.810310 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a5cf24b4-dd05-45a4-bf8f-5895c30ada04-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"a5cf24b4-dd05-45a4-bf8f-5895c30ada04\") " pod="openstack/ovn-northd-0" Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.810336 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a5cf24b4-dd05-45a4-bf8f-5895c30ada04-scripts\") pod \"ovn-northd-0\" (UID: \"a5cf24b4-dd05-45a4-bf8f-5895c30ada04\") " pod="openstack/ovn-northd-0" Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.870403 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-7xdqf" event={"ID":"93fb09e4-51b3-44f7-afab-c49c374c7f56","Type":"ContainerStarted","Data":"c22452b770dd4a9a65921ab0ffd83302ac4e18edaf96f19a7ef93fd9aed1ddbd"} Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.883676 4873 generic.go:334] "Generic (PLEG): container finished" podID="d5588863-2c50-4860-ac45-4d226f59f4ab" containerID="99d10df78458a4b918d3e158949e5fd8107d001f23798facc6d8904c218ffc35" exitCode=0 Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.883773 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-zng4c" event={"ID":"d5588863-2c50-4860-ac45-4d226f59f4ab","Type":"ContainerDied","Data":"99d10df78458a4b918d3e158949e5fd8107d001f23798facc6d8904c218ffc35"} Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.894804 4873 generic.go:334] "Generic (PLEG): container finished" podID="34cbfd49-8920-4490-9f8a-7a6f2542d1c6" containerID="5a1f8f98f1c2680b702cd2ff1d7aedc07c140dd8eda4a5cc515f58974915bc70" exitCode=0 Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.894899 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-gpkrq" event={"ID":"34cbfd49-8920-4490-9f8a-7a6f2542d1c6","Type":"ContainerDied","Data":"5a1f8f98f1c2680b702cd2ff1d7aedc07c140dd8eda4a5cc515f58974915bc70"} Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.899405 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"250cbd69-c69a-475c-8973-bc98caf4b264","Type":"ContainerStarted","Data":"2c8fdd4e738442a3be36ca1141b6df4236314a2f0e6ac64a598629ab30c8be07"} Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.912932 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5cf24b4-dd05-45a4-bf8f-5895c30ada04-config\") pod \"ovn-northd-0\" (UID: \"a5cf24b4-dd05-45a4-bf8f-5895c30ada04\") " pod="openstack/ovn-northd-0" Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.912985 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5cf24b4-dd05-45a4-bf8f-5895c30ada04-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"a5cf24b4-dd05-45a4-bf8f-5895c30ada04\") " pod="openstack/ovn-northd-0" Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.913043 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5cf24b4-dd05-45a4-bf8f-5895c30ada04-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"a5cf24b4-dd05-45a4-bf8f-5895c30ada04\") " pod="openstack/ovn-northd-0" Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.913102 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cf24b4-dd05-45a4-bf8f-5895c30ada04-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"a5cf24b4-dd05-45a4-bf8f-5895c30ada04\") " pod="openstack/ovn-northd-0" Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.913129 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fv2jn\" (UniqueName: \"kubernetes.io/projected/a5cf24b4-dd05-45a4-bf8f-5895c30ada04-kube-api-access-fv2jn\") pod \"ovn-northd-0\" (UID: \"a5cf24b4-dd05-45a4-bf8f-5895c30ada04\") " pod="openstack/ovn-northd-0" Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.913152 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a5cf24b4-dd05-45a4-bf8f-5895c30ada04-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"a5cf24b4-dd05-45a4-bf8f-5895c30ada04\") " pod="openstack/ovn-northd-0" Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.913173 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a5cf24b4-dd05-45a4-bf8f-5895c30ada04-scripts\") pod \"ovn-northd-0\" (UID: \"a5cf24b4-dd05-45a4-bf8f-5895c30ada04\") " pod="openstack/ovn-northd-0" Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.914559 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5cf24b4-dd05-45a4-bf8f-5895c30ada04-config\") pod \"ovn-northd-0\" (UID: \"a5cf24b4-dd05-45a4-bf8f-5895c30ada04\") " pod="openstack/ovn-northd-0" Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.915763 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a5cf24b4-dd05-45a4-bf8f-5895c30ada04-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"a5cf24b4-dd05-45a4-bf8f-5895c30ada04\") " pod="openstack/ovn-northd-0" Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.921234 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cf24b4-dd05-45a4-bf8f-5895c30ada04-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"a5cf24b4-dd05-45a4-bf8f-5895c30ada04\") " pod="openstack/ovn-northd-0" Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.922297 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a5cf24b4-dd05-45a4-bf8f-5895c30ada04-scripts\") pod \"ovn-northd-0\" (UID: \"a5cf24b4-dd05-45a4-bf8f-5895c30ada04\") " pod="openstack/ovn-northd-0" Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.929024 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5cf24b4-dd05-45a4-bf8f-5895c30ada04-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"a5cf24b4-dd05-45a4-bf8f-5895c30ada04\") " pod="openstack/ovn-northd-0" Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.972141 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fv2jn\" (UniqueName: \"kubernetes.io/projected/a5cf24b4-dd05-45a4-bf8f-5895c30ada04-kube-api-access-fv2jn\") pod \"ovn-northd-0\" (UID: \"a5cf24b4-dd05-45a4-bf8f-5895c30ada04\") " pod="openstack/ovn-northd-0" Dec 01 08:58:06 crc kubenswrapper[4873]: I1201 08:58:06.981612 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5cf24b4-dd05-45a4-bf8f-5895c30ada04-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"a5cf24b4-dd05-45a4-bf8f-5895c30ada04\") " pod="openstack/ovn-northd-0" Dec 01 08:58:07 crc kubenswrapper[4873]: I1201 08:58:07.038150 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 01 08:58:07 crc kubenswrapper[4873]: I1201 08:58:07.063153 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=35.462021661 podStartE2EDuration="43.063129644s" podCreationTimestamp="2025-12-01 08:57:24 +0000 UTC" firstStartedPulling="2025-12-01 08:57:50.770746963 +0000 UTC m=+1046.672855502" lastFinishedPulling="2025-12-01 08:57:58.371854946 +0000 UTC m=+1054.273963485" observedRunningTime="2025-12-01 08:58:06.952497545 +0000 UTC m=+1062.854606084" watchObservedRunningTime="2025-12-01 08:58:07.063129644 +0000 UTC m=+1062.965238183" Dec 01 08:58:07 crc kubenswrapper[4873]: I1201 08:58:07.066099 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-x2trf"] Dec 01 08:58:07 crc kubenswrapper[4873]: W1201 08:58:07.092556 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod35124c22_9257_440f_8a97_9e029187bd69.slice/crio-81626255da32a61663eb180e49811136f37e584801fb586f317e25bde1e630a6 WatchSource:0}: Error finding container 81626255da32a61663eb180e49811136f37e584801fb586f317e25bde1e630a6: Status 404 returned error can't find the container with id 81626255da32a61663eb180e49811136f37e584801fb586f317e25bde1e630a6 Dec 01 08:58:07 crc kubenswrapper[4873]: I1201 08:58:07.128332 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-gpkrq" Dec 01 08:58:07 crc kubenswrapper[4873]: I1201 08:58:07.291710 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 01 08:58:07 crc kubenswrapper[4873]: I1201 08:58:07.294519 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 01 08:58:07 crc kubenswrapper[4873]: I1201 08:58:07.348504 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-j7n86"] Dec 01 08:58:07 crc kubenswrapper[4873]: I1201 08:58:07.357723 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-zng4c" Dec 01 08:58:07 crc kubenswrapper[4873]: I1201 08:58:07.426151 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d5588863-2c50-4860-ac45-4d226f59f4ab-dns-svc\") pod \"d5588863-2c50-4860-ac45-4d226f59f4ab\" (UID: \"d5588863-2c50-4860-ac45-4d226f59f4ab\") " Dec 01 08:58:07 crc kubenswrapper[4873]: I1201 08:58:07.426421 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34cbfd49-8920-4490-9f8a-7a6f2542d1c6-config\") pod \"34cbfd49-8920-4490-9f8a-7a6f2542d1c6\" (UID: \"34cbfd49-8920-4490-9f8a-7a6f2542d1c6\") " Dec 01 08:58:07 crc kubenswrapper[4873]: I1201 08:58:07.426468 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5588863-2c50-4860-ac45-4d226f59f4ab-config\") pod \"d5588863-2c50-4860-ac45-4d226f59f4ab\" (UID: \"d5588863-2c50-4860-ac45-4d226f59f4ab\") " Dec 01 08:58:07 crc kubenswrapper[4873]: I1201 08:58:07.426513 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-smnfg\" (UniqueName: \"kubernetes.io/projected/d5588863-2c50-4860-ac45-4d226f59f4ab-kube-api-access-smnfg\") pod \"d5588863-2c50-4860-ac45-4d226f59f4ab\" (UID: \"d5588863-2c50-4860-ac45-4d226f59f4ab\") " Dec 01 08:58:07 crc kubenswrapper[4873]: I1201 08:58:07.427945 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5tzzd\" (UniqueName: \"kubernetes.io/projected/34cbfd49-8920-4490-9f8a-7a6f2542d1c6-kube-api-access-5tzzd\") pod \"34cbfd49-8920-4490-9f8a-7a6f2542d1c6\" (UID: \"34cbfd49-8920-4490-9f8a-7a6f2542d1c6\") " Dec 01 08:58:07 crc kubenswrapper[4873]: I1201 08:58:07.428191 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/34cbfd49-8920-4490-9f8a-7a6f2542d1c6-dns-svc\") pod \"34cbfd49-8920-4490-9f8a-7a6f2542d1c6\" (UID: \"34cbfd49-8920-4490-9f8a-7a6f2542d1c6\") " Dec 01 08:58:07 crc kubenswrapper[4873]: I1201 08:58:07.441505 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5588863-2c50-4860-ac45-4d226f59f4ab-kube-api-access-smnfg" (OuterVolumeSpecName: "kube-api-access-smnfg") pod "d5588863-2c50-4860-ac45-4d226f59f4ab" (UID: "d5588863-2c50-4860-ac45-4d226f59f4ab"). InnerVolumeSpecName "kube-api-access-smnfg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:58:07 crc kubenswrapper[4873]: I1201 08:58:07.447121 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34cbfd49-8920-4490-9f8a-7a6f2542d1c6-kube-api-access-5tzzd" (OuterVolumeSpecName: "kube-api-access-5tzzd") pod "34cbfd49-8920-4490-9f8a-7a6f2542d1c6" (UID: "34cbfd49-8920-4490-9f8a-7a6f2542d1c6"). InnerVolumeSpecName "kube-api-access-5tzzd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:58:07 crc kubenswrapper[4873]: I1201 08:58:07.482088 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5588863-2c50-4860-ac45-4d226f59f4ab-config" (OuterVolumeSpecName: "config") pod "d5588863-2c50-4860-ac45-4d226f59f4ab" (UID: "d5588863-2c50-4860-ac45-4d226f59f4ab"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:58:07 crc kubenswrapper[4873]: I1201 08:58:07.483243 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34cbfd49-8920-4490-9f8a-7a6f2542d1c6-config" (OuterVolumeSpecName: "config") pod "34cbfd49-8920-4490-9f8a-7a6f2542d1c6" (UID: "34cbfd49-8920-4490-9f8a-7a6f2542d1c6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:58:07 crc kubenswrapper[4873]: I1201 08:58:07.532401 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5tzzd\" (UniqueName: \"kubernetes.io/projected/34cbfd49-8920-4490-9f8a-7a6f2542d1c6-kube-api-access-5tzzd\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:07 crc kubenswrapper[4873]: I1201 08:58:07.532724 4873 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34cbfd49-8920-4490-9f8a-7a6f2542d1c6-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:07 crc kubenswrapper[4873]: I1201 08:58:07.532736 4873 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5588863-2c50-4860-ac45-4d226f59f4ab-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:07 crc kubenswrapper[4873]: I1201 08:58:07.532748 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-smnfg\" (UniqueName: \"kubernetes.io/projected/d5588863-2c50-4860-ac45-4d226f59f4ab-kube-api-access-smnfg\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:07 crc kubenswrapper[4873]: I1201 08:58:07.553287 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5588863-2c50-4860-ac45-4d226f59f4ab-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d5588863-2c50-4860-ac45-4d226f59f4ab" (UID: "d5588863-2c50-4860-ac45-4d226f59f4ab"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:58:07 crc kubenswrapper[4873]: I1201 08:58:07.584169 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34cbfd49-8920-4490-9f8a-7a6f2542d1c6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "34cbfd49-8920-4490-9f8a-7a6f2542d1c6" (UID: "34cbfd49-8920-4490-9f8a-7a6f2542d1c6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:58:07 crc kubenswrapper[4873]: I1201 08:58:07.640648 4873 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d5588863-2c50-4860-ac45-4d226f59f4ab-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:07 crc kubenswrapper[4873]: I1201 08:58:07.640682 4873 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/34cbfd49-8920-4490-9f8a-7a6f2542d1c6-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:07 crc kubenswrapper[4873]: I1201 08:58:07.653369 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 01 08:58:07 crc kubenswrapper[4873]: I1201 08:58:07.725396 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 01 08:58:07 crc kubenswrapper[4873]: W1201 08:58:07.753043 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda5cf24b4_dd05_45a4_bf8f_5895c30ada04.slice/crio-7cb21274ad0d8810d90d40636f2a1d08ce98d7b61a8d3d8a592220f10170737b WatchSource:0}: Error finding container 7cb21274ad0d8810d90d40636f2a1d08ce98d7b61a8d3d8a592220f10170737b: Status 404 returned error can't find the container with id 7cb21274ad0d8810d90d40636f2a1d08ce98d7b61a8d3d8a592220f10170737b Dec 01 08:58:07 crc kubenswrapper[4873]: I1201 08:58:07.910718 4873 generic.go:334] "Generic (PLEG): container finished" podID="35124c22-9257-440f-8a97-9e029187bd69" containerID="55da1e6165b9660dcc9e19ab827260d62e16d9ed64f17c5331875d6696ad2a3b" exitCode=0 Dec 01 08:58:07 crc kubenswrapper[4873]: I1201 08:58:07.910819 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-x2trf" event={"ID":"35124c22-9257-440f-8a97-9e029187bd69","Type":"ContainerDied","Data":"55da1e6165b9660dcc9e19ab827260d62e16d9ed64f17c5331875d6696ad2a3b"} Dec 01 08:58:07 crc kubenswrapper[4873]: I1201 08:58:07.911974 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-x2trf" event={"ID":"35124c22-9257-440f-8a97-9e029187bd69","Type":"ContainerStarted","Data":"81626255da32a61663eb180e49811136f37e584801fb586f317e25bde1e630a6"} Dec 01 08:58:07 crc kubenswrapper[4873]: I1201 08:58:07.918073 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-gpkrq" event={"ID":"34cbfd49-8920-4490-9f8a-7a6f2542d1c6","Type":"ContainerDied","Data":"33f8d169979df01a49e2cc720645b8540426da2ba6056033c1226b45556d345c"} Dec 01 08:58:07 crc kubenswrapper[4873]: I1201 08:58:07.918141 4873 scope.go:117] "RemoveContainer" containerID="5a1f8f98f1c2680b702cd2ff1d7aedc07c140dd8eda4a5cc515f58974915bc70" Dec 01 08:58:07 crc kubenswrapper[4873]: I1201 08:58:07.918203 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-gpkrq" Dec 01 08:58:07 crc kubenswrapper[4873]: I1201 08:58:07.920610 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-7xdqf" event={"ID":"93fb09e4-51b3-44f7-afab-c49c374c7f56","Type":"ContainerStarted","Data":"f9f07181662b7d7e5e03c12138de49853a8177a8f2571325535c7139d4ec08ee"} Dec 01 08:58:07 crc kubenswrapper[4873]: I1201 08:58:07.921932 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a5cf24b4-dd05-45a4-bf8f-5895c30ada04","Type":"ContainerStarted","Data":"7cb21274ad0d8810d90d40636f2a1d08ce98d7b61a8d3d8a592220f10170737b"} Dec 01 08:58:07 crc kubenswrapper[4873]: I1201 08:58:07.923729 4873 generic.go:334] "Generic (PLEG): container finished" podID="11ec3abe-6052-445d-9f3f-016157fecbf9" containerID="f3643b9dfddb59b751277cd5d49925dbeae682a21e8f4202abe8de17ee25358e" exitCode=0 Dec 01 08:58:07 crc kubenswrapper[4873]: I1201 08:58:07.924691 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-j7n86" event={"ID":"11ec3abe-6052-445d-9f3f-016157fecbf9","Type":"ContainerDied","Data":"f3643b9dfddb59b751277cd5d49925dbeae682a21e8f4202abe8de17ee25358e"} Dec 01 08:58:07 crc kubenswrapper[4873]: I1201 08:58:07.924841 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-j7n86" event={"ID":"11ec3abe-6052-445d-9f3f-016157fecbf9","Type":"ContainerStarted","Data":"28c37e91ea5eba47df338112136f5e1f44d396525851c5242ebca47b9ce4addd"} Dec 01 08:58:07 crc kubenswrapper[4873]: I1201 08:58:07.928284 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-zng4c" event={"ID":"d5588863-2c50-4860-ac45-4d226f59f4ab","Type":"ContainerDied","Data":"7cbaa3ed8027f8a8dd10a930e57faa3ae56cff6b435a1f8f62351abfdf311464"} Dec 01 08:58:07 crc kubenswrapper[4873]: I1201 08:58:07.928326 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-zng4c" Dec 01 08:58:08 crc kubenswrapper[4873]: I1201 08:58:08.005546 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-7xdqf" podStartSLOduration=3.005516954 podStartE2EDuration="3.005516954s" podCreationTimestamp="2025-12-01 08:58:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:58:08.003578289 +0000 UTC m=+1063.905686828" watchObservedRunningTime="2025-12-01 08:58:08.005516954 +0000 UTC m=+1063.907625493" Dec 01 08:58:08 crc kubenswrapper[4873]: I1201 08:58:08.096622 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-gpkrq"] Dec 01 08:58:08 crc kubenswrapper[4873]: I1201 08:58:08.116220 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-gpkrq"] Dec 01 08:58:08 crc kubenswrapper[4873]: I1201 08:58:08.121182 4873 scope.go:117] "RemoveContainer" containerID="1a8c12fd81b743f8d8e4e6b676148b2a4c7f62c53371bb478d28fc5af2831d19" Dec 01 08:58:08 crc kubenswrapper[4873]: I1201 08:58:08.123364 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-zng4c"] Dec 01 08:58:08 crc kubenswrapper[4873]: I1201 08:58:08.129622 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-zng4c"] Dec 01 08:58:08 crc kubenswrapper[4873]: I1201 08:58:08.148658 4873 scope.go:117] "RemoveContainer" containerID="99d10df78458a4b918d3e158949e5fd8107d001f23798facc6d8904c218ffc35" Dec 01 08:58:08 crc kubenswrapper[4873]: I1201 08:58:08.184949 4873 scope.go:117] "RemoveContainer" containerID="16ad9f3f5dbafb64151afe4ad7fd52500cb0482d633e2933de657e75ecc5105f" Dec 01 08:58:08 crc kubenswrapper[4873]: I1201 08:58:08.442970 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34cbfd49-8920-4490-9f8a-7a6f2542d1c6" path="/var/lib/kubelet/pods/34cbfd49-8920-4490-9f8a-7a6f2542d1c6/volumes" Dec 01 08:58:08 crc kubenswrapper[4873]: I1201 08:58:08.443686 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5588863-2c50-4860-ac45-4d226f59f4ab" path="/var/lib/kubelet/pods/d5588863-2c50-4860-ac45-4d226f59f4ab/volumes" Dec 01 08:58:08 crc kubenswrapper[4873]: I1201 08:58:08.992345 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-j7n86" event={"ID":"11ec3abe-6052-445d-9f3f-016157fecbf9","Type":"ContainerStarted","Data":"e8b70798f88aed17ff235aba1f24cb3657753f831be3e65ea9bb413ec4f1b38e"} Dec 01 08:58:08 crc kubenswrapper[4873]: I1201 08:58:08.992701 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-j7n86" Dec 01 08:58:08 crc kubenswrapper[4873]: I1201 08:58:08.999467 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-x2trf" event={"ID":"35124c22-9257-440f-8a97-9e029187bd69","Type":"ContainerStarted","Data":"4d835e5b44dbd3d0d53cdddde1da0e965a0a778af89b624c78179472424c2348"} Dec 01 08:58:08 crc kubenswrapper[4873]: I1201 08:58:08.999581 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7f896c8c65-x2trf" Dec 01 08:58:09 crc kubenswrapper[4873]: I1201 08:58:09.014751 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-j7n86" podStartSLOduration=3.014719199 podStartE2EDuration="3.014719199s" podCreationTimestamp="2025-12-01 08:58:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:58:09.011602916 +0000 UTC m=+1064.913711455" watchObservedRunningTime="2025-12-01 08:58:09.014719199 +0000 UTC m=+1064.916827738" Dec 01 08:58:09 crc kubenswrapper[4873]: I1201 08:58:09.035553 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7f896c8c65-x2trf" podStartSLOduration=4.03552746 podStartE2EDuration="4.03552746s" podCreationTimestamp="2025-12-01 08:58:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:58:09.032533849 +0000 UTC m=+1064.934642398" watchObservedRunningTime="2025-12-01 08:58:09.03552746 +0000 UTC m=+1064.937635999" Dec 01 08:58:09 crc kubenswrapper[4873]: I1201 08:58:09.311411 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 01 08:58:10 crc kubenswrapper[4873]: I1201 08:58:10.024899 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a5cf24b4-dd05-45a4-bf8f-5895c30ada04","Type":"ContainerStarted","Data":"0a39e685fb9029dac037bb079302a130f5487b7a20f6a48b12b597988e1ef5ba"} Dec 01 08:58:11 crc kubenswrapper[4873]: I1201 08:58:11.036137 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a5cf24b4-dd05-45a4-bf8f-5895c30ada04","Type":"ContainerStarted","Data":"27c04b6beb6f73a50f591abed68b7f5051da2ad0ba83a42c89bc64c50fc8cb28"} Dec 01 08:58:11 crc kubenswrapper[4873]: I1201 08:58:11.037735 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 01 08:58:11 crc kubenswrapper[4873]: I1201 08:58:11.073861 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=3.505439772 podStartE2EDuration="5.073831211s" podCreationTimestamp="2025-12-01 08:58:06 +0000 UTC" firstStartedPulling="2025-12-01 08:58:07.771166389 +0000 UTC m=+1063.673274928" lastFinishedPulling="2025-12-01 08:58:09.339557828 +0000 UTC m=+1065.241666367" observedRunningTime="2025-12-01 08:58:11.064159803 +0000 UTC m=+1066.966268352" watchObservedRunningTime="2025-12-01 08:58:11.073831211 +0000 UTC m=+1066.975939750" Dec 01 08:58:15 crc kubenswrapper[4873]: I1201 08:58:15.480695 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 01 08:58:15 crc kubenswrapper[4873]: I1201 08:58:15.583788 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 01 08:58:15 crc kubenswrapper[4873]: I1201 08:58:15.982300 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7f896c8c65-x2trf" Dec 01 08:58:16 crc kubenswrapper[4873]: I1201 08:58:16.165868 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 01 08:58:16 crc kubenswrapper[4873]: I1201 08:58:16.165978 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 01 08:58:16 crc kubenswrapper[4873]: I1201 08:58:16.448457 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 01 08:58:16 crc kubenswrapper[4873]: I1201 08:58:16.459228 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86db49b7ff-j7n86" Dec 01 08:58:16 crc kubenswrapper[4873]: I1201 08:58:16.555881 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-x2trf"] Dec 01 08:58:16 crc kubenswrapper[4873]: I1201 08:58:16.557503 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7f896c8c65-x2trf" podUID="35124c22-9257-440f-8a97-9e029187bd69" containerName="dnsmasq-dns" containerID="cri-o://4d835e5b44dbd3d0d53cdddde1da0e965a0a778af89b624c78179472424c2348" gracePeriod=10 Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.108664 4873 generic.go:334] "Generic (PLEG): container finished" podID="35124c22-9257-440f-8a97-9e029187bd69" containerID="4d835e5b44dbd3d0d53cdddde1da0e965a0a778af89b624c78179472424c2348" exitCode=0 Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.109622 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-x2trf" event={"ID":"35124c22-9257-440f-8a97-9e029187bd69","Type":"ContainerDied","Data":"4d835e5b44dbd3d0d53cdddde1da0e965a0a778af89b624c78179472424c2348"} Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.205251 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.603680 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-x2trf" Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.652506 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-8kwtw"] Dec 01 08:58:17 crc kubenswrapper[4873]: E1201 08:58:17.653038 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34cbfd49-8920-4490-9f8a-7a6f2542d1c6" containerName="init" Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.653058 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="34cbfd49-8920-4490-9f8a-7a6f2542d1c6" containerName="init" Dec 01 08:58:17 crc kubenswrapper[4873]: E1201 08:58:17.653069 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35124c22-9257-440f-8a97-9e029187bd69" containerName="dnsmasq-dns" Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.653076 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="35124c22-9257-440f-8a97-9e029187bd69" containerName="dnsmasq-dns" Dec 01 08:58:17 crc kubenswrapper[4873]: E1201 08:58:17.653086 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5588863-2c50-4860-ac45-4d226f59f4ab" containerName="init" Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.653092 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5588863-2c50-4860-ac45-4d226f59f4ab" containerName="init" Dec 01 08:58:17 crc kubenswrapper[4873]: E1201 08:58:17.653110 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34cbfd49-8920-4490-9f8a-7a6f2542d1c6" containerName="dnsmasq-dns" Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.653117 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="34cbfd49-8920-4490-9f8a-7a6f2542d1c6" containerName="dnsmasq-dns" Dec 01 08:58:17 crc kubenswrapper[4873]: E1201 08:58:17.653129 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5588863-2c50-4860-ac45-4d226f59f4ab" containerName="dnsmasq-dns" Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.653135 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5588863-2c50-4860-ac45-4d226f59f4ab" containerName="dnsmasq-dns" Dec 01 08:58:17 crc kubenswrapper[4873]: E1201 08:58:17.653153 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35124c22-9257-440f-8a97-9e029187bd69" containerName="init" Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.653159 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="35124c22-9257-440f-8a97-9e029187bd69" containerName="init" Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.653355 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5588863-2c50-4860-ac45-4d226f59f4ab" containerName="dnsmasq-dns" Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.653373 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="34cbfd49-8920-4490-9f8a-7a6f2542d1c6" containerName="dnsmasq-dns" Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.653388 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="35124c22-9257-440f-8a97-9e029187bd69" containerName="dnsmasq-dns" Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.654175 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-8kwtw" Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.662143 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-1033-account-create-update-mfx62"] Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.663685 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-1033-account-create-update-mfx62" Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.667595 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.690177 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-8kwtw"] Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.697814 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-1033-account-create-update-mfx62"] Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.700480 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35124c22-9257-440f-8a97-9e029187bd69-config\") pod \"35124c22-9257-440f-8a97-9e029187bd69\" (UID: \"35124c22-9257-440f-8a97-9e029187bd69\") " Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.700573 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35124c22-9257-440f-8a97-9e029187bd69-dns-svc\") pod \"35124c22-9257-440f-8a97-9e029187bd69\" (UID: \"35124c22-9257-440f-8a97-9e029187bd69\") " Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.700745 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35124c22-9257-440f-8a97-9e029187bd69-ovsdbserver-sb\") pod \"35124c22-9257-440f-8a97-9e029187bd69\" (UID: \"35124c22-9257-440f-8a97-9e029187bd69\") " Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.700822 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4449\" (UniqueName: \"kubernetes.io/projected/35124c22-9257-440f-8a97-9e029187bd69-kube-api-access-f4449\") pod \"35124c22-9257-440f-8a97-9e029187bd69\" (UID: \"35124c22-9257-440f-8a97-9e029187bd69\") " Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.701182 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8bc6fcca-1230-4aee-aee3-169b34e8c830-operator-scripts\") pod \"placement-db-create-8kwtw\" (UID: \"8bc6fcca-1230-4aee-aee3-169b34e8c830\") " pod="openstack/placement-db-create-8kwtw" Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.701291 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hfp8\" (UniqueName: \"kubernetes.io/projected/00b50e4b-c0ee-4186-95d1-3d793632ceff-kube-api-access-5hfp8\") pod \"placement-1033-account-create-update-mfx62\" (UID: \"00b50e4b-c0ee-4186-95d1-3d793632ceff\") " pod="openstack/placement-1033-account-create-update-mfx62" Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.701346 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00b50e4b-c0ee-4186-95d1-3d793632ceff-operator-scripts\") pod \"placement-1033-account-create-update-mfx62\" (UID: \"00b50e4b-c0ee-4186-95d1-3d793632ceff\") " pod="openstack/placement-1033-account-create-update-mfx62" Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.701410 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5dt2\" (UniqueName: \"kubernetes.io/projected/8bc6fcca-1230-4aee-aee3-169b34e8c830-kube-api-access-x5dt2\") pod \"placement-db-create-8kwtw\" (UID: \"8bc6fcca-1230-4aee-aee3-169b34e8c830\") " pod="openstack/placement-db-create-8kwtw" Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.712170 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35124c22-9257-440f-8a97-9e029187bd69-kube-api-access-f4449" (OuterVolumeSpecName: "kube-api-access-f4449") pod "35124c22-9257-440f-8a97-9e029187bd69" (UID: "35124c22-9257-440f-8a97-9e029187bd69"). InnerVolumeSpecName "kube-api-access-f4449". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.749761 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35124c22-9257-440f-8a97-9e029187bd69-config" (OuterVolumeSpecName: "config") pod "35124c22-9257-440f-8a97-9e029187bd69" (UID: "35124c22-9257-440f-8a97-9e029187bd69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.778252 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35124c22-9257-440f-8a97-9e029187bd69-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "35124c22-9257-440f-8a97-9e029187bd69" (UID: "35124c22-9257-440f-8a97-9e029187bd69"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.778754 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35124c22-9257-440f-8a97-9e029187bd69-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "35124c22-9257-440f-8a97-9e029187bd69" (UID: "35124c22-9257-440f-8a97-9e029187bd69"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.803618 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8bc6fcca-1230-4aee-aee3-169b34e8c830-operator-scripts\") pod \"placement-db-create-8kwtw\" (UID: \"8bc6fcca-1230-4aee-aee3-169b34e8c830\") " pod="openstack/placement-db-create-8kwtw" Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.803748 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hfp8\" (UniqueName: \"kubernetes.io/projected/00b50e4b-c0ee-4186-95d1-3d793632ceff-kube-api-access-5hfp8\") pod \"placement-1033-account-create-update-mfx62\" (UID: \"00b50e4b-c0ee-4186-95d1-3d793632ceff\") " pod="openstack/placement-1033-account-create-update-mfx62" Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.803799 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00b50e4b-c0ee-4186-95d1-3d793632ceff-operator-scripts\") pod \"placement-1033-account-create-update-mfx62\" (UID: \"00b50e4b-c0ee-4186-95d1-3d793632ceff\") " pod="openstack/placement-1033-account-create-update-mfx62" Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.803852 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5dt2\" (UniqueName: \"kubernetes.io/projected/8bc6fcca-1230-4aee-aee3-169b34e8c830-kube-api-access-x5dt2\") pod \"placement-db-create-8kwtw\" (UID: \"8bc6fcca-1230-4aee-aee3-169b34e8c830\") " pod="openstack/placement-db-create-8kwtw" Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.803940 4873 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35124c22-9257-440f-8a97-9e029187bd69-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.803957 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4449\" (UniqueName: \"kubernetes.io/projected/35124c22-9257-440f-8a97-9e029187bd69-kube-api-access-f4449\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.803975 4873 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35124c22-9257-440f-8a97-9e029187bd69-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.803990 4873 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35124c22-9257-440f-8a97-9e029187bd69-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.804926 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8bc6fcca-1230-4aee-aee3-169b34e8c830-operator-scripts\") pod \"placement-db-create-8kwtw\" (UID: \"8bc6fcca-1230-4aee-aee3-169b34e8c830\") " pod="openstack/placement-db-create-8kwtw" Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.805746 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00b50e4b-c0ee-4186-95d1-3d793632ceff-operator-scripts\") pod \"placement-1033-account-create-update-mfx62\" (UID: \"00b50e4b-c0ee-4186-95d1-3d793632ceff\") " pod="openstack/placement-1033-account-create-update-mfx62" Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.827724 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hfp8\" (UniqueName: \"kubernetes.io/projected/00b50e4b-c0ee-4186-95d1-3d793632ceff-kube-api-access-5hfp8\") pod \"placement-1033-account-create-update-mfx62\" (UID: \"00b50e4b-c0ee-4186-95d1-3d793632ceff\") " pod="openstack/placement-1033-account-create-update-mfx62" Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.832383 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5dt2\" (UniqueName: \"kubernetes.io/projected/8bc6fcca-1230-4aee-aee3-169b34e8c830-kube-api-access-x5dt2\") pod \"placement-db-create-8kwtw\" (UID: \"8bc6fcca-1230-4aee-aee3-169b34e8c830\") " pod="openstack/placement-db-create-8kwtw" Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.845494 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-4p8hz"] Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.846873 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-4p8hz" Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.861902 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-4p8hz"] Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.906488 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4r867\" (UniqueName: \"kubernetes.io/projected/5ca75802-f944-4266-80b0-3599a0c0311d-kube-api-access-4r867\") pod \"glance-db-create-4p8hz\" (UID: \"5ca75802-f944-4266-80b0-3599a0c0311d\") " pod="openstack/glance-db-create-4p8hz" Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.906998 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ca75802-f944-4266-80b0-3599a0c0311d-operator-scripts\") pod \"glance-db-create-4p8hz\" (UID: \"5ca75802-f944-4266-80b0-3599a0c0311d\") " pod="openstack/glance-db-create-4p8hz" Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.950066 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-23e6-account-create-update-2h4sc"] Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.951542 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-23e6-account-create-update-2h4sc" Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.953909 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.969044 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-23e6-account-create-update-2h4sc"] Dec 01 08:58:17 crc kubenswrapper[4873]: I1201 08:58:17.987932 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-1033-account-create-update-mfx62" Dec 01 08:58:18 crc kubenswrapper[4873]: I1201 08:58:18.009583 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4r867\" (UniqueName: \"kubernetes.io/projected/5ca75802-f944-4266-80b0-3599a0c0311d-kube-api-access-4r867\") pod \"glance-db-create-4p8hz\" (UID: \"5ca75802-f944-4266-80b0-3599a0c0311d\") " pod="openstack/glance-db-create-4p8hz" Dec 01 08:58:18 crc kubenswrapper[4873]: I1201 08:58:18.009750 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ca75802-f944-4266-80b0-3599a0c0311d-operator-scripts\") pod \"glance-db-create-4p8hz\" (UID: \"5ca75802-f944-4266-80b0-3599a0c0311d\") " pod="openstack/glance-db-create-4p8hz" Dec 01 08:58:18 crc kubenswrapper[4873]: I1201 08:58:18.011199 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ca75802-f944-4266-80b0-3599a0c0311d-operator-scripts\") pod \"glance-db-create-4p8hz\" (UID: \"5ca75802-f944-4266-80b0-3599a0c0311d\") " pod="openstack/glance-db-create-4p8hz" Dec 01 08:58:18 crc kubenswrapper[4873]: I1201 08:58:18.016580 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-8kwtw" Dec 01 08:58:18 crc kubenswrapper[4873]: I1201 08:58:18.041714 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4r867\" (UniqueName: \"kubernetes.io/projected/5ca75802-f944-4266-80b0-3599a0c0311d-kube-api-access-4r867\") pod \"glance-db-create-4p8hz\" (UID: \"5ca75802-f944-4266-80b0-3599a0c0311d\") " pod="openstack/glance-db-create-4p8hz" Dec 01 08:58:18 crc kubenswrapper[4873]: I1201 08:58:18.111226 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d995b8fc-09c8-4ac2-9387-a6ead74e1571-operator-scripts\") pod \"glance-23e6-account-create-update-2h4sc\" (UID: \"d995b8fc-09c8-4ac2-9387-a6ead74e1571\") " pod="openstack/glance-23e6-account-create-update-2h4sc" Dec 01 08:58:18 crc kubenswrapper[4873]: I1201 08:58:18.111276 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwccl\" (UniqueName: \"kubernetes.io/projected/d995b8fc-09c8-4ac2-9387-a6ead74e1571-kube-api-access-hwccl\") pod \"glance-23e6-account-create-update-2h4sc\" (UID: \"d995b8fc-09c8-4ac2-9387-a6ead74e1571\") " pod="openstack/glance-23e6-account-create-update-2h4sc" Dec 01 08:58:18 crc kubenswrapper[4873]: I1201 08:58:18.123217 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-x2trf" event={"ID":"35124c22-9257-440f-8a97-9e029187bd69","Type":"ContainerDied","Data":"81626255da32a61663eb180e49811136f37e584801fb586f317e25bde1e630a6"} Dec 01 08:58:18 crc kubenswrapper[4873]: I1201 08:58:18.123278 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-x2trf" Dec 01 08:58:18 crc kubenswrapper[4873]: I1201 08:58:18.123307 4873 scope.go:117] "RemoveContainer" containerID="4d835e5b44dbd3d0d53cdddde1da0e965a0a778af89b624c78179472424c2348" Dec 01 08:58:18 crc kubenswrapper[4873]: I1201 08:58:18.148716 4873 scope.go:117] "RemoveContainer" containerID="55da1e6165b9660dcc9e19ab827260d62e16d9ed64f17c5331875d6696ad2a3b" Dec 01 08:58:18 crc kubenswrapper[4873]: I1201 08:58:18.181837 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-x2trf"] Dec 01 08:58:18 crc kubenswrapper[4873]: I1201 08:58:18.191583 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-x2trf"] Dec 01 08:58:18 crc kubenswrapper[4873]: I1201 08:58:18.215543 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d995b8fc-09c8-4ac2-9387-a6ead74e1571-operator-scripts\") pod \"glance-23e6-account-create-update-2h4sc\" (UID: \"d995b8fc-09c8-4ac2-9387-a6ead74e1571\") " pod="openstack/glance-23e6-account-create-update-2h4sc" Dec 01 08:58:18 crc kubenswrapper[4873]: I1201 08:58:18.215609 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwccl\" (UniqueName: \"kubernetes.io/projected/d995b8fc-09c8-4ac2-9387-a6ead74e1571-kube-api-access-hwccl\") pod \"glance-23e6-account-create-update-2h4sc\" (UID: \"d995b8fc-09c8-4ac2-9387-a6ead74e1571\") " pod="openstack/glance-23e6-account-create-update-2h4sc" Dec 01 08:58:18 crc kubenswrapper[4873]: I1201 08:58:18.217134 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d995b8fc-09c8-4ac2-9387-a6ead74e1571-operator-scripts\") pod \"glance-23e6-account-create-update-2h4sc\" (UID: \"d995b8fc-09c8-4ac2-9387-a6ead74e1571\") " pod="openstack/glance-23e6-account-create-update-2h4sc" Dec 01 08:58:18 crc kubenswrapper[4873]: I1201 08:58:18.222731 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-4p8hz" Dec 01 08:58:18 crc kubenswrapper[4873]: I1201 08:58:18.236730 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwccl\" (UniqueName: \"kubernetes.io/projected/d995b8fc-09c8-4ac2-9387-a6ead74e1571-kube-api-access-hwccl\") pod \"glance-23e6-account-create-update-2h4sc\" (UID: \"d995b8fc-09c8-4ac2-9387-a6ead74e1571\") " pod="openstack/glance-23e6-account-create-update-2h4sc" Dec 01 08:58:18 crc kubenswrapper[4873]: I1201 08:58:18.331918 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-23e6-account-create-update-2h4sc" Dec 01 08:58:18 crc kubenswrapper[4873]: I1201 08:58:18.446921 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35124c22-9257-440f-8a97-9e029187bd69" path="/var/lib/kubelet/pods/35124c22-9257-440f-8a97-9e029187bd69/volumes" Dec 01 08:58:18 crc kubenswrapper[4873]: I1201 08:58:18.551955 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-1033-account-create-update-mfx62"] Dec 01 08:58:18 crc kubenswrapper[4873]: I1201 08:58:18.717442 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-4p8hz"] Dec 01 08:58:18 crc kubenswrapper[4873]: I1201 08:58:18.817920 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-23e6-account-create-update-2h4sc"] Dec 01 08:58:18 crc kubenswrapper[4873]: I1201 08:58:18.900393 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-8kwtw"] Dec 01 08:58:19 crc kubenswrapper[4873]: I1201 08:58:19.141392 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-8kwtw" event={"ID":"8bc6fcca-1230-4aee-aee3-169b34e8c830","Type":"ContainerStarted","Data":"a4f4d85b4304f7ddd266a8231c61ca5b48da15ceee230462125f7d3587e33e0a"} Dec 01 08:58:19 crc kubenswrapper[4873]: I1201 08:58:19.143494 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-4p8hz" event={"ID":"5ca75802-f944-4266-80b0-3599a0c0311d","Type":"ContainerStarted","Data":"aba12e9adedc8a0feede329d963df752cb4abb4aac5b21dbfd923e7b47ae44a6"} Dec 01 08:58:19 crc kubenswrapper[4873]: I1201 08:58:19.143536 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-4p8hz" event={"ID":"5ca75802-f944-4266-80b0-3599a0c0311d","Type":"ContainerStarted","Data":"0133f3901382c7aaa166112ef64496f2e31d56f134f4b7f0056bbbf1db55f0dc"} Dec 01 08:58:19 crc kubenswrapper[4873]: I1201 08:58:19.145620 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-1033-account-create-update-mfx62" event={"ID":"00b50e4b-c0ee-4186-95d1-3d793632ceff","Type":"ContainerStarted","Data":"c71e7e8cd5db62b0e94da0859a60c4145cf609266c32f087c66fe0a385c1874b"} Dec 01 08:58:19 crc kubenswrapper[4873]: I1201 08:58:19.145652 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-1033-account-create-update-mfx62" event={"ID":"00b50e4b-c0ee-4186-95d1-3d793632ceff","Type":"ContainerStarted","Data":"bae4ce03c08dab45659f375045b72f9dc21254999955bde6e2bb579e5ce8c37e"} Dec 01 08:58:19 crc kubenswrapper[4873]: I1201 08:58:19.150581 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-23e6-account-create-update-2h4sc" event={"ID":"d995b8fc-09c8-4ac2-9387-a6ead74e1571","Type":"ContainerStarted","Data":"aae082fdae3fc250879341c001c1761664636ec139d30fb2c1e83a050ed0d960"} Dec 01 08:58:19 crc kubenswrapper[4873]: I1201 08:58:19.175048 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-4p8hz" podStartSLOduration=2.175006835 podStartE2EDuration="2.175006835s" podCreationTimestamp="2025-12-01 08:58:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:58:19.168780299 +0000 UTC m=+1075.070888838" watchObservedRunningTime="2025-12-01 08:58:19.175006835 +0000 UTC m=+1075.077115374" Dec 01 08:58:19 crc kubenswrapper[4873]: I1201 08:58:19.199147 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-1033-account-create-update-mfx62" podStartSLOduration=2.199124484 podStartE2EDuration="2.199124484s" podCreationTimestamp="2025-12-01 08:58:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:58:19.193048521 +0000 UTC m=+1075.095157060" watchObservedRunningTime="2025-12-01 08:58:19.199124484 +0000 UTC m=+1075.101233023" Dec 01 08:58:20 crc kubenswrapper[4873]: I1201 08:58:20.160103 4873 generic.go:334] "Generic (PLEG): container finished" podID="d995b8fc-09c8-4ac2-9387-a6ead74e1571" containerID="93054c797b00a9970940d90143c89b4861d57853b5dd9fab3bc751985d960276" exitCode=0 Dec 01 08:58:20 crc kubenswrapper[4873]: I1201 08:58:20.160190 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-23e6-account-create-update-2h4sc" event={"ID":"d995b8fc-09c8-4ac2-9387-a6ead74e1571","Type":"ContainerDied","Data":"93054c797b00a9970940d90143c89b4861d57853b5dd9fab3bc751985d960276"} Dec 01 08:58:20 crc kubenswrapper[4873]: I1201 08:58:20.162901 4873 generic.go:334] "Generic (PLEG): container finished" podID="8bc6fcca-1230-4aee-aee3-169b34e8c830" containerID="9f7e8b7218c9ef0fe039ad1d0baf4dd982f364a3a9363b0368d64452fc74200f" exitCode=0 Dec 01 08:58:20 crc kubenswrapper[4873]: I1201 08:58:20.162968 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-8kwtw" event={"ID":"8bc6fcca-1230-4aee-aee3-169b34e8c830","Type":"ContainerDied","Data":"9f7e8b7218c9ef0fe039ad1d0baf4dd982f364a3a9363b0368d64452fc74200f"} Dec 01 08:58:20 crc kubenswrapper[4873]: I1201 08:58:20.165257 4873 generic.go:334] "Generic (PLEG): container finished" podID="5ca75802-f944-4266-80b0-3599a0c0311d" containerID="aba12e9adedc8a0feede329d963df752cb4abb4aac5b21dbfd923e7b47ae44a6" exitCode=0 Dec 01 08:58:20 crc kubenswrapper[4873]: I1201 08:58:20.165334 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-4p8hz" event={"ID":"5ca75802-f944-4266-80b0-3599a0c0311d","Type":"ContainerDied","Data":"aba12e9adedc8a0feede329d963df752cb4abb4aac5b21dbfd923e7b47ae44a6"} Dec 01 08:58:20 crc kubenswrapper[4873]: I1201 08:58:20.167303 4873 generic.go:334] "Generic (PLEG): container finished" podID="00b50e4b-c0ee-4186-95d1-3d793632ceff" containerID="c71e7e8cd5db62b0e94da0859a60c4145cf609266c32f087c66fe0a385c1874b" exitCode=0 Dec 01 08:58:20 crc kubenswrapper[4873]: I1201 08:58:20.167366 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-1033-account-create-update-mfx62" event={"ID":"00b50e4b-c0ee-4186-95d1-3d793632ceff","Type":"ContainerDied","Data":"c71e7e8cd5db62b0e94da0859a60c4145cf609266c32f087c66fe0a385c1874b"} Dec 01 08:58:21 crc kubenswrapper[4873]: I1201 08:58:21.651133 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-1033-account-create-update-mfx62" Dec 01 08:58:21 crc kubenswrapper[4873]: I1201 08:58:21.710404 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00b50e4b-c0ee-4186-95d1-3d793632ceff-operator-scripts\") pod \"00b50e4b-c0ee-4186-95d1-3d793632ceff\" (UID: \"00b50e4b-c0ee-4186-95d1-3d793632ceff\") " Dec 01 08:58:21 crc kubenswrapper[4873]: I1201 08:58:21.710665 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5hfp8\" (UniqueName: \"kubernetes.io/projected/00b50e4b-c0ee-4186-95d1-3d793632ceff-kube-api-access-5hfp8\") pod \"00b50e4b-c0ee-4186-95d1-3d793632ceff\" (UID: \"00b50e4b-c0ee-4186-95d1-3d793632ceff\") " Dec 01 08:58:21 crc kubenswrapper[4873]: I1201 08:58:21.712709 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00b50e4b-c0ee-4186-95d1-3d793632ceff-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "00b50e4b-c0ee-4186-95d1-3d793632ceff" (UID: "00b50e4b-c0ee-4186-95d1-3d793632ceff"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:58:21 crc kubenswrapper[4873]: I1201 08:58:21.728924 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00b50e4b-c0ee-4186-95d1-3d793632ceff-kube-api-access-5hfp8" (OuterVolumeSpecName: "kube-api-access-5hfp8") pod "00b50e4b-c0ee-4186-95d1-3d793632ceff" (UID: "00b50e4b-c0ee-4186-95d1-3d793632ceff"). InnerVolumeSpecName "kube-api-access-5hfp8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:58:21 crc kubenswrapper[4873]: I1201 08:58:21.813432 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5hfp8\" (UniqueName: \"kubernetes.io/projected/00b50e4b-c0ee-4186-95d1-3d793632ceff-kube-api-access-5hfp8\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:21 crc kubenswrapper[4873]: I1201 08:58:21.813492 4873 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00b50e4b-c0ee-4186-95d1-3d793632ceff-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:21 crc kubenswrapper[4873]: I1201 08:58:21.847546 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-8kwtw" Dec 01 08:58:21 crc kubenswrapper[4873]: I1201 08:58:21.856548 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-4p8hz" Dec 01 08:58:21 crc kubenswrapper[4873]: I1201 08:58:21.875577 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-23e6-account-create-update-2h4sc" Dec 01 08:58:21 crc kubenswrapper[4873]: I1201 08:58:21.915419 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4r867\" (UniqueName: \"kubernetes.io/projected/5ca75802-f944-4266-80b0-3599a0c0311d-kube-api-access-4r867\") pod \"5ca75802-f944-4266-80b0-3599a0c0311d\" (UID: \"5ca75802-f944-4266-80b0-3599a0c0311d\") " Dec 01 08:58:21 crc kubenswrapper[4873]: I1201 08:58:21.915600 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5dt2\" (UniqueName: \"kubernetes.io/projected/8bc6fcca-1230-4aee-aee3-169b34e8c830-kube-api-access-x5dt2\") pod \"8bc6fcca-1230-4aee-aee3-169b34e8c830\" (UID: \"8bc6fcca-1230-4aee-aee3-169b34e8c830\") " Dec 01 08:58:21 crc kubenswrapper[4873]: I1201 08:58:21.915658 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ca75802-f944-4266-80b0-3599a0c0311d-operator-scripts\") pod \"5ca75802-f944-4266-80b0-3599a0c0311d\" (UID: \"5ca75802-f944-4266-80b0-3599a0c0311d\") " Dec 01 08:58:21 crc kubenswrapper[4873]: I1201 08:58:21.915838 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d995b8fc-09c8-4ac2-9387-a6ead74e1571-operator-scripts\") pod \"d995b8fc-09c8-4ac2-9387-a6ead74e1571\" (UID: \"d995b8fc-09c8-4ac2-9387-a6ead74e1571\") " Dec 01 08:58:21 crc kubenswrapper[4873]: I1201 08:58:21.915945 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8bc6fcca-1230-4aee-aee3-169b34e8c830-operator-scripts\") pod \"8bc6fcca-1230-4aee-aee3-169b34e8c830\" (UID: \"8bc6fcca-1230-4aee-aee3-169b34e8c830\") " Dec 01 08:58:21 crc kubenswrapper[4873]: I1201 08:58:21.916027 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hwccl\" (UniqueName: \"kubernetes.io/projected/d995b8fc-09c8-4ac2-9387-a6ead74e1571-kube-api-access-hwccl\") pod \"d995b8fc-09c8-4ac2-9387-a6ead74e1571\" (UID: \"d995b8fc-09c8-4ac2-9387-a6ead74e1571\") " Dec 01 08:58:21 crc kubenswrapper[4873]: I1201 08:58:21.918556 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ca75802-f944-4266-80b0-3599a0c0311d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5ca75802-f944-4266-80b0-3599a0c0311d" (UID: "5ca75802-f944-4266-80b0-3599a0c0311d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:58:21 crc kubenswrapper[4873]: I1201 08:58:21.919572 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8bc6fcca-1230-4aee-aee3-169b34e8c830-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8bc6fcca-1230-4aee-aee3-169b34e8c830" (UID: "8bc6fcca-1230-4aee-aee3-169b34e8c830"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:58:21 crc kubenswrapper[4873]: I1201 08:58:21.920119 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d995b8fc-09c8-4ac2-9387-a6ead74e1571-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d995b8fc-09c8-4ac2-9387-a6ead74e1571" (UID: "d995b8fc-09c8-4ac2-9387-a6ead74e1571"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:58:21 crc kubenswrapper[4873]: I1201 08:58:21.923295 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d995b8fc-09c8-4ac2-9387-a6ead74e1571-kube-api-access-hwccl" (OuterVolumeSpecName: "kube-api-access-hwccl") pod "d995b8fc-09c8-4ac2-9387-a6ead74e1571" (UID: "d995b8fc-09c8-4ac2-9387-a6ead74e1571"). InnerVolumeSpecName "kube-api-access-hwccl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:58:21 crc kubenswrapper[4873]: I1201 08:58:21.925690 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ca75802-f944-4266-80b0-3599a0c0311d-kube-api-access-4r867" (OuterVolumeSpecName: "kube-api-access-4r867") pod "5ca75802-f944-4266-80b0-3599a0c0311d" (UID: "5ca75802-f944-4266-80b0-3599a0c0311d"). InnerVolumeSpecName "kube-api-access-4r867". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:58:21 crc kubenswrapper[4873]: I1201 08:58:21.927268 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bc6fcca-1230-4aee-aee3-169b34e8c830-kube-api-access-x5dt2" (OuterVolumeSpecName: "kube-api-access-x5dt2") pod "8bc6fcca-1230-4aee-aee3-169b34e8c830" (UID: "8bc6fcca-1230-4aee-aee3-169b34e8c830"). InnerVolumeSpecName "kube-api-access-x5dt2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:58:22 crc kubenswrapper[4873]: I1201 08:58:22.019030 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5dt2\" (UniqueName: \"kubernetes.io/projected/8bc6fcca-1230-4aee-aee3-169b34e8c830-kube-api-access-x5dt2\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:22 crc kubenswrapper[4873]: I1201 08:58:22.019082 4873 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ca75802-f944-4266-80b0-3599a0c0311d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:22 crc kubenswrapper[4873]: I1201 08:58:22.019095 4873 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d995b8fc-09c8-4ac2-9387-a6ead74e1571-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:22 crc kubenswrapper[4873]: I1201 08:58:22.019106 4873 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8bc6fcca-1230-4aee-aee3-169b34e8c830-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:22 crc kubenswrapper[4873]: I1201 08:58:22.019116 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hwccl\" (UniqueName: \"kubernetes.io/projected/d995b8fc-09c8-4ac2-9387-a6ead74e1571-kube-api-access-hwccl\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:22 crc kubenswrapper[4873]: I1201 08:58:22.019125 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4r867\" (UniqueName: \"kubernetes.io/projected/5ca75802-f944-4266-80b0-3599a0c0311d-kube-api-access-4r867\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:22 crc kubenswrapper[4873]: I1201 08:58:22.108936 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 01 08:58:22 crc kubenswrapper[4873]: I1201 08:58:22.184572 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-1033-account-create-update-mfx62" event={"ID":"00b50e4b-c0ee-4186-95d1-3d793632ceff","Type":"ContainerDied","Data":"bae4ce03c08dab45659f375045b72f9dc21254999955bde6e2bb579e5ce8c37e"} Dec 01 08:58:22 crc kubenswrapper[4873]: I1201 08:58:22.184983 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bae4ce03c08dab45659f375045b72f9dc21254999955bde6e2bb579e5ce8c37e" Dec 01 08:58:22 crc kubenswrapper[4873]: I1201 08:58:22.185124 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-1033-account-create-update-mfx62" Dec 01 08:58:22 crc kubenswrapper[4873]: I1201 08:58:22.188718 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-23e6-account-create-update-2h4sc" event={"ID":"d995b8fc-09c8-4ac2-9387-a6ead74e1571","Type":"ContainerDied","Data":"aae082fdae3fc250879341c001c1761664636ec139d30fb2c1e83a050ed0d960"} Dec 01 08:58:22 crc kubenswrapper[4873]: I1201 08:58:22.188773 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aae082fdae3fc250879341c001c1761664636ec139d30fb2c1e83a050ed0d960" Dec 01 08:58:22 crc kubenswrapper[4873]: I1201 08:58:22.188845 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-23e6-account-create-update-2h4sc" Dec 01 08:58:22 crc kubenswrapper[4873]: I1201 08:58:22.192483 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-8kwtw" Dec 01 08:58:22 crc kubenswrapper[4873]: I1201 08:58:22.192505 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-8kwtw" event={"ID":"8bc6fcca-1230-4aee-aee3-169b34e8c830","Type":"ContainerDied","Data":"a4f4d85b4304f7ddd266a8231c61ca5b48da15ceee230462125f7d3587e33e0a"} Dec 01 08:58:22 crc kubenswrapper[4873]: I1201 08:58:22.192573 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a4f4d85b4304f7ddd266a8231c61ca5b48da15ceee230462125f7d3587e33e0a" Dec 01 08:58:22 crc kubenswrapper[4873]: I1201 08:58:22.194168 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-4p8hz" event={"ID":"5ca75802-f944-4266-80b0-3599a0c0311d","Type":"ContainerDied","Data":"0133f3901382c7aaa166112ef64496f2e31d56f134f4b7f0056bbbf1db55f0dc"} Dec 01 08:58:22 crc kubenswrapper[4873]: I1201 08:58:22.194229 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0133f3901382c7aaa166112ef64496f2e31d56f134f4b7f0056bbbf1db55f0dc" Dec 01 08:58:22 crc kubenswrapper[4873]: I1201 08:58:22.194322 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-4p8hz" Dec 01 08:58:23 crc kubenswrapper[4873]: I1201 08:58:23.234935 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-6k8d4"] Dec 01 08:58:23 crc kubenswrapper[4873]: E1201 08:58:23.236255 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ca75802-f944-4266-80b0-3599a0c0311d" containerName="mariadb-database-create" Dec 01 08:58:23 crc kubenswrapper[4873]: I1201 08:58:23.236285 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ca75802-f944-4266-80b0-3599a0c0311d" containerName="mariadb-database-create" Dec 01 08:58:23 crc kubenswrapper[4873]: E1201 08:58:23.236308 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d995b8fc-09c8-4ac2-9387-a6ead74e1571" containerName="mariadb-account-create-update" Dec 01 08:58:23 crc kubenswrapper[4873]: I1201 08:58:23.236317 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="d995b8fc-09c8-4ac2-9387-a6ead74e1571" containerName="mariadb-account-create-update" Dec 01 08:58:23 crc kubenswrapper[4873]: E1201 08:58:23.236332 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bc6fcca-1230-4aee-aee3-169b34e8c830" containerName="mariadb-database-create" Dec 01 08:58:23 crc kubenswrapper[4873]: I1201 08:58:23.236341 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bc6fcca-1230-4aee-aee3-169b34e8c830" containerName="mariadb-database-create" Dec 01 08:58:23 crc kubenswrapper[4873]: E1201 08:58:23.236353 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00b50e4b-c0ee-4186-95d1-3d793632ceff" containerName="mariadb-account-create-update" Dec 01 08:58:23 crc kubenswrapper[4873]: I1201 08:58:23.236360 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="00b50e4b-c0ee-4186-95d1-3d793632ceff" containerName="mariadb-account-create-update" Dec 01 08:58:23 crc kubenswrapper[4873]: I1201 08:58:23.236580 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="00b50e4b-c0ee-4186-95d1-3d793632ceff" containerName="mariadb-account-create-update" Dec 01 08:58:23 crc kubenswrapper[4873]: I1201 08:58:23.236605 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ca75802-f944-4266-80b0-3599a0c0311d" containerName="mariadb-database-create" Dec 01 08:58:23 crc kubenswrapper[4873]: I1201 08:58:23.236617 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bc6fcca-1230-4aee-aee3-169b34e8c830" containerName="mariadb-database-create" Dec 01 08:58:23 crc kubenswrapper[4873]: I1201 08:58:23.236627 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="d995b8fc-09c8-4ac2-9387-a6ead74e1571" containerName="mariadb-account-create-update" Dec 01 08:58:23 crc kubenswrapper[4873]: I1201 08:58:23.237553 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-6k8d4" Dec 01 08:58:23 crc kubenswrapper[4873]: I1201 08:58:23.240434 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 01 08:58:23 crc kubenswrapper[4873]: I1201 08:58:23.249539 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-jmjhl" Dec 01 08:58:23 crc kubenswrapper[4873]: I1201 08:58:23.252338 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-6k8d4"] Dec 01 08:58:23 crc kubenswrapper[4873]: I1201 08:58:23.343738 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8-config-data\") pod \"glance-db-sync-6k8d4\" (UID: \"0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8\") " pod="openstack/glance-db-sync-6k8d4" Dec 01 08:58:23 crc kubenswrapper[4873]: I1201 08:58:23.343915 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8-combined-ca-bundle\") pod \"glance-db-sync-6k8d4\" (UID: \"0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8\") " pod="openstack/glance-db-sync-6k8d4" Dec 01 08:58:23 crc kubenswrapper[4873]: I1201 08:58:23.344041 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5q8q\" (UniqueName: \"kubernetes.io/projected/0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8-kube-api-access-s5q8q\") pod \"glance-db-sync-6k8d4\" (UID: \"0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8\") " pod="openstack/glance-db-sync-6k8d4" Dec 01 08:58:23 crc kubenswrapper[4873]: I1201 08:58:23.344347 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8-db-sync-config-data\") pod \"glance-db-sync-6k8d4\" (UID: \"0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8\") " pod="openstack/glance-db-sync-6k8d4" Dec 01 08:58:23 crc kubenswrapper[4873]: I1201 08:58:23.501486 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8-db-sync-config-data\") pod \"glance-db-sync-6k8d4\" (UID: \"0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8\") " pod="openstack/glance-db-sync-6k8d4" Dec 01 08:58:23 crc kubenswrapper[4873]: I1201 08:58:23.501940 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8-config-data\") pod \"glance-db-sync-6k8d4\" (UID: \"0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8\") " pod="openstack/glance-db-sync-6k8d4" Dec 01 08:58:23 crc kubenswrapper[4873]: I1201 08:58:23.502331 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8-combined-ca-bundle\") pod \"glance-db-sync-6k8d4\" (UID: \"0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8\") " pod="openstack/glance-db-sync-6k8d4" Dec 01 08:58:23 crc kubenswrapper[4873]: I1201 08:58:23.502393 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5q8q\" (UniqueName: \"kubernetes.io/projected/0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8-kube-api-access-s5q8q\") pod \"glance-db-sync-6k8d4\" (UID: \"0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8\") " pod="openstack/glance-db-sync-6k8d4" Dec 01 08:58:23 crc kubenswrapper[4873]: I1201 08:58:23.510977 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8-combined-ca-bundle\") pod \"glance-db-sync-6k8d4\" (UID: \"0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8\") " pod="openstack/glance-db-sync-6k8d4" Dec 01 08:58:23 crc kubenswrapper[4873]: I1201 08:58:23.513705 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8-config-data\") pod \"glance-db-sync-6k8d4\" (UID: \"0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8\") " pod="openstack/glance-db-sync-6k8d4" Dec 01 08:58:23 crc kubenswrapper[4873]: I1201 08:58:23.516625 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8-db-sync-config-data\") pod \"glance-db-sync-6k8d4\" (UID: \"0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8\") " pod="openstack/glance-db-sync-6k8d4" Dec 01 08:58:23 crc kubenswrapper[4873]: I1201 08:58:23.525120 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5q8q\" (UniqueName: \"kubernetes.io/projected/0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8-kube-api-access-s5q8q\") pod \"glance-db-sync-6k8d4\" (UID: \"0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8\") " pod="openstack/glance-db-sync-6k8d4" Dec 01 08:58:23 crc kubenswrapper[4873]: I1201 08:58:23.569609 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-6k8d4" Dec 01 08:58:24 crc kubenswrapper[4873]: I1201 08:58:24.001684 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-6k8d4"] Dec 01 08:58:24 crc kubenswrapper[4873]: I1201 08:58:24.216962 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-6k8d4" event={"ID":"0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8","Type":"ContainerStarted","Data":"d6aaf3fd77cfca960bc6f24e53269da0172d1489fa9b03a5ba2c73d244f3bff1"} Dec 01 08:58:27 crc kubenswrapper[4873]: I1201 08:58:27.333426 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-552e-account-create-update-9r7p2"] Dec 01 08:58:27 crc kubenswrapper[4873]: I1201 08:58:27.335415 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-552e-account-create-update-9r7p2" Dec 01 08:58:27 crc kubenswrapper[4873]: I1201 08:58:27.339119 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 01 08:58:27 crc kubenswrapper[4873]: I1201 08:58:27.349053 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-552e-account-create-update-9r7p2"] Dec 01 08:58:27 crc kubenswrapper[4873]: I1201 08:58:27.382873 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98acc7b6-efd7-49b8-bcd5-ae1d67848dd7-operator-scripts\") pod \"keystone-552e-account-create-update-9r7p2\" (UID: \"98acc7b6-efd7-49b8-bcd5-ae1d67848dd7\") " pod="openstack/keystone-552e-account-create-update-9r7p2" Dec 01 08:58:27 crc kubenswrapper[4873]: I1201 08:58:27.382990 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68gd9\" (UniqueName: \"kubernetes.io/projected/98acc7b6-efd7-49b8-bcd5-ae1d67848dd7-kube-api-access-68gd9\") pod \"keystone-552e-account-create-update-9r7p2\" (UID: \"98acc7b6-efd7-49b8-bcd5-ae1d67848dd7\") " pod="openstack/keystone-552e-account-create-update-9r7p2" Dec 01 08:58:27 crc kubenswrapper[4873]: I1201 08:58:27.432006 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-cbzpc"] Dec 01 08:58:27 crc kubenswrapper[4873]: I1201 08:58:27.433687 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-cbzpc" Dec 01 08:58:27 crc kubenswrapper[4873]: I1201 08:58:27.453856 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-cbzpc"] Dec 01 08:58:27 crc kubenswrapper[4873]: I1201 08:58:27.485200 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98acc7b6-efd7-49b8-bcd5-ae1d67848dd7-operator-scripts\") pod \"keystone-552e-account-create-update-9r7p2\" (UID: \"98acc7b6-efd7-49b8-bcd5-ae1d67848dd7\") " pod="openstack/keystone-552e-account-create-update-9r7p2" Dec 01 08:58:27 crc kubenswrapper[4873]: I1201 08:58:27.486887 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98acc7b6-efd7-49b8-bcd5-ae1d67848dd7-operator-scripts\") pod \"keystone-552e-account-create-update-9r7p2\" (UID: \"98acc7b6-efd7-49b8-bcd5-ae1d67848dd7\") " pod="openstack/keystone-552e-account-create-update-9r7p2" Dec 01 08:58:27 crc kubenswrapper[4873]: I1201 08:58:27.486920 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68gd9\" (UniqueName: \"kubernetes.io/projected/98acc7b6-efd7-49b8-bcd5-ae1d67848dd7-kube-api-access-68gd9\") pod \"keystone-552e-account-create-update-9r7p2\" (UID: \"98acc7b6-efd7-49b8-bcd5-ae1d67848dd7\") " pod="openstack/keystone-552e-account-create-update-9r7p2" Dec 01 08:58:27 crc kubenswrapper[4873]: I1201 08:58:27.518874 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68gd9\" (UniqueName: \"kubernetes.io/projected/98acc7b6-efd7-49b8-bcd5-ae1d67848dd7-kube-api-access-68gd9\") pod \"keystone-552e-account-create-update-9r7p2\" (UID: \"98acc7b6-efd7-49b8-bcd5-ae1d67848dd7\") " pod="openstack/keystone-552e-account-create-update-9r7p2" Dec 01 08:58:27 crc kubenswrapper[4873]: I1201 08:58:27.588858 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d1118a3-3429-457f-bfcf-40e9f173798e-operator-scripts\") pod \"keystone-db-create-cbzpc\" (UID: \"9d1118a3-3429-457f-bfcf-40e9f173798e\") " pod="openstack/keystone-db-create-cbzpc" Dec 01 08:58:27 crc kubenswrapper[4873]: I1201 08:58:27.589610 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ld7sv\" (UniqueName: \"kubernetes.io/projected/9d1118a3-3429-457f-bfcf-40e9f173798e-kube-api-access-ld7sv\") pod \"keystone-db-create-cbzpc\" (UID: \"9d1118a3-3429-457f-bfcf-40e9f173798e\") " pod="openstack/keystone-db-create-cbzpc" Dec 01 08:58:27 crc kubenswrapper[4873]: I1201 08:58:27.656318 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-552e-account-create-update-9r7p2" Dec 01 08:58:27 crc kubenswrapper[4873]: I1201 08:58:27.692131 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ld7sv\" (UniqueName: \"kubernetes.io/projected/9d1118a3-3429-457f-bfcf-40e9f173798e-kube-api-access-ld7sv\") pod \"keystone-db-create-cbzpc\" (UID: \"9d1118a3-3429-457f-bfcf-40e9f173798e\") " pod="openstack/keystone-db-create-cbzpc" Dec 01 08:58:27 crc kubenswrapper[4873]: I1201 08:58:27.692631 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d1118a3-3429-457f-bfcf-40e9f173798e-operator-scripts\") pod \"keystone-db-create-cbzpc\" (UID: \"9d1118a3-3429-457f-bfcf-40e9f173798e\") " pod="openstack/keystone-db-create-cbzpc" Dec 01 08:58:27 crc kubenswrapper[4873]: I1201 08:58:27.693920 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d1118a3-3429-457f-bfcf-40e9f173798e-operator-scripts\") pod \"keystone-db-create-cbzpc\" (UID: \"9d1118a3-3429-457f-bfcf-40e9f173798e\") " pod="openstack/keystone-db-create-cbzpc" Dec 01 08:58:27 crc kubenswrapper[4873]: I1201 08:58:27.716899 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ld7sv\" (UniqueName: \"kubernetes.io/projected/9d1118a3-3429-457f-bfcf-40e9f173798e-kube-api-access-ld7sv\") pod \"keystone-db-create-cbzpc\" (UID: \"9d1118a3-3429-457f-bfcf-40e9f173798e\") " pod="openstack/keystone-db-create-cbzpc" Dec 01 08:58:27 crc kubenswrapper[4873]: I1201 08:58:27.755587 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-cbzpc" Dec 01 08:58:28 crc kubenswrapper[4873]: I1201 08:58:28.482090 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-552e-account-create-update-9r7p2"] Dec 01 08:58:28 crc kubenswrapper[4873]: I1201 08:58:28.501031 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-cbzpc"] Dec 01 08:58:28 crc kubenswrapper[4873]: W1201 08:58:28.518370 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d1118a3_3429_457f_bfcf_40e9f173798e.slice/crio-333c58f68b5094cc4b4efecc77944b615d7eeb3bcc75a0d8084b460616d57439 WatchSource:0}: Error finding container 333c58f68b5094cc4b4efecc77944b615d7eeb3bcc75a0d8084b460616d57439: Status 404 returned error can't find the container with id 333c58f68b5094cc4b4efecc77944b615d7eeb3bcc75a0d8084b460616d57439 Dec 01 08:58:29 crc kubenswrapper[4873]: I1201 08:58:29.298410 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-cbzpc" event={"ID":"9d1118a3-3429-457f-bfcf-40e9f173798e","Type":"ContainerStarted","Data":"333c58f68b5094cc4b4efecc77944b615d7eeb3bcc75a0d8084b460616d57439"} Dec 01 08:58:29 crc kubenswrapper[4873]: I1201 08:58:29.300346 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-552e-account-create-update-9r7p2" event={"ID":"98acc7b6-efd7-49b8-bcd5-ae1d67848dd7","Type":"ContainerStarted","Data":"a26163b58ee631f42e044c0b6f4b6045df6fb317856cd049caf0a2ca725c0a49"} Dec 01 08:58:30 crc kubenswrapper[4873]: I1201 08:58:30.323600 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-552e-account-create-update-9r7p2" event={"ID":"98acc7b6-efd7-49b8-bcd5-ae1d67848dd7","Type":"ContainerStarted","Data":"229af48591ddaf7172ced7c2664e0ea50f54009b4e96c2798791a8ac56514807"} Dec 01 08:58:31 crc kubenswrapper[4873]: I1201 08:58:31.059809 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:58:31 crc kubenswrapper[4873]: I1201 08:58:31.059945 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:58:31 crc kubenswrapper[4873]: I1201 08:58:31.339810 4873 generic.go:334] "Generic (PLEG): container finished" podID="9d1118a3-3429-457f-bfcf-40e9f173798e" containerID="924d488dca72675a7c8014797ad15892332b3f100666693db69c08f4fb13fdc0" exitCode=0 Dec 01 08:58:31 crc kubenswrapper[4873]: I1201 08:58:31.339901 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-cbzpc" event={"ID":"9d1118a3-3429-457f-bfcf-40e9f173798e","Type":"ContainerDied","Data":"924d488dca72675a7c8014797ad15892332b3f100666693db69c08f4fb13fdc0"} Dec 01 08:58:31 crc kubenswrapper[4873]: I1201 08:58:31.342585 4873 generic.go:334] "Generic (PLEG): container finished" podID="98acc7b6-efd7-49b8-bcd5-ae1d67848dd7" containerID="229af48591ddaf7172ced7c2664e0ea50f54009b4e96c2798791a8ac56514807" exitCode=0 Dec 01 08:58:31 crc kubenswrapper[4873]: I1201 08:58:31.342618 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-552e-account-create-update-9r7p2" event={"ID":"98acc7b6-efd7-49b8-bcd5-ae1d67848dd7","Type":"ContainerDied","Data":"229af48591ddaf7172ced7c2664e0ea50f54009b4e96c2798791a8ac56514807"} Dec 01 08:58:33 crc kubenswrapper[4873]: I1201 08:58:33.330776 4873 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-w8v6j" podUID="b79364c5-00b2-4ab8-b2f0-aff17ed902b4" containerName="ovn-controller" probeResult="failure" output=< Dec 01 08:58:33 crc kubenswrapper[4873]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 01 08:58:33 crc kubenswrapper[4873]: > Dec 01 08:58:33 crc kubenswrapper[4873]: I1201 08:58:33.406992 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-xd66n" Dec 01 08:58:33 crc kubenswrapper[4873]: I1201 08:58:33.414767 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-xd66n" Dec 01 08:58:33 crc kubenswrapper[4873]: I1201 08:58:33.676822 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-w8v6j-config-w58gr"] Dec 01 08:58:33 crc kubenswrapper[4873]: I1201 08:58:33.681962 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-w8v6j-config-w58gr" Dec 01 08:58:33 crc kubenswrapper[4873]: I1201 08:58:33.689301 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 01 08:58:33 crc kubenswrapper[4873]: I1201 08:58:33.702135 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-w8v6j-config-w58gr"] Dec 01 08:58:33 crc kubenswrapper[4873]: I1201 08:58:33.741638 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6ae1598f-7693-4112-bc8d-a46e3c5e7f57-var-run-ovn\") pod \"ovn-controller-w8v6j-config-w58gr\" (UID: \"6ae1598f-7693-4112-bc8d-a46e3c5e7f57\") " pod="openstack/ovn-controller-w8v6j-config-w58gr" Dec 01 08:58:33 crc kubenswrapper[4873]: I1201 08:58:33.741707 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6ae1598f-7693-4112-bc8d-a46e3c5e7f57-var-log-ovn\") pod \"ovn-controller-w8v6j-config-w58gr\" (UID: \"6ae1598f-7693-4112-bc8d-a46e3c5e7f57\") " pod="openstack/ovn-controller-w8v6j-config-w58gr" Dec 01 08:58:33 crc kubenswrapper[4873]: I1201 08:58:33.741753 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6ae1598f-7693-4112-bc8d-a46e3c5e7f57-var-run\") pod \"ovn-controller-w8v6j-config-w58gr\" (UID: \"6ae1598f-7693-4112-bc8d-a46e3c5e7f57\") " pod="openstack/ovn-controller-w8v6j-config-w58gr" Dec 01 08:58:33 crc kubenswrapper[4873]: I1201 08:58:33.741789 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6ae1598f-7693-4112-bc8d-a46e3c5e7f57-scripts\") pod \"ovn-controller-w8v6j-config-w58gr\" (UID: \"6ae1598f-7693-4112-bc8d-a46e3c5e7f57\") " pod="openstack/ovn-controller-w8v6j-config-w58gr" Dec 01 08:58:33 crc kubenswrapper[4873]: I1201 08:58:33.741820 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6ae1598f-7693-4112-bc8d-a46e3c5e7f57-additional-scripts\") pod \"ovn-controller-w8v6j-config-w58gr\" (UID: \"6ae1598f-7693-4112-bc8d-a46e3c5e7f57\") " pod="openstack/ovn-controller-w8v6j-config-w58gr" Dec 01 08:58:33 crc kubenswrapper[4873]: I1201 08:58:33.741850 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dw6t\" (UniqueName: \"kubernetes.io/projected/6ae1598f-7693-4112-bc8d-a46e3c5e7f57-kube-api-access-5dw6t\") pod \"ovn-controller-w8v6j-config-w58gr\" (UID: \"6ae1598f-7693-4112-bc8d-a46e3c5e7f57\") " pod="openstack/ovn-controller-w8v6j-config-w58gr" Dec 01 08:58:33 crc kubenswrapper[4873]: I1201 08:58:33.843960 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6ae1598f-7693-4112-bc8d-a46e3c5e7f57-var-run-ovn\") pod \"ovn-controller-w8v6j-config-w58gr\" (UID: \"6ae1598f-7693-4112-bc8d-a46e3c5e7f57\") " pod="openstack/ovn-controller-w8v6j-config-w58gr" Dec 01 08:58:33 crc kubenswrapper[4873]: I1201 08:58:33.844047 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6ae1598f-7693-4112-bc8d-a46e3c5e7f57-var-log-ovn\") pod \"ovn-controller-w8v6j-config-w58gr\" (UID: \"6ae1598f-7693-4112-bc8d-a46e3c5e7f57\") " pod="openstack/ovn-controller-w8v6j-config-w58gr" Dec 01 08:58:33 crc kubenswrapper[4873]: I1201 08:58:33.844091 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6ae1598f-7693-4112-bc8d-a46e3c5e7f57-var-run\") pod \"ovn-controller-w8v6j-config-w58gr\" (UID: \"6ae1598f-7693-4112-bc8d-a46e3c5e7f57\") " pod="openstack/ovn-controller-w8v6j-config-w58gr" Dec 01 08:58:33 crc kubenswrapper[4873]: I1201 08:58:33.844141 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6ae1598f-7693-4112-bc8d-a46e3c5e7f57-scripts\") pod \"ovn-controller-w8v6j-config-w58gr\" (UID: \"6ae1598f-7693-4112-bc8d-a46e3c5e7f57\") " pod="openstack/ovn-controller-w8v6j-config-w58gr" Dec 01 08:58:33 crc kubenswrapper[4873]: I1201 08:58:33.844183 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6ae1598f-7693-4112-bc8d-a46e3c5e7f57-additional-scripts\") pod \"ovn-controller-w8v6j-config-w58gr\" (UID: \"6ae1598f-7693-4112-bc8d-a46e3c5e7f57\") " pod="openstack/ovn-controller-w8v6j-config-w58gr" Dec 01 08:58:33 crc kubenswrapper[4873]: I1201 08:58:33.844221 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dw6t\" (UniqueName: \"kubernetes.io/projected/6ae1598f-7693-4112-bc8d-a46e3c5e7f57-kube-api-access-5dw6t\") pod \"ovn-controller-w8v6j-config-w58gr\" (UID: \"6ae1598f-7693-4112-bc8d-a46e3c5e7f57\") " pod="openstack/ovn-controller-w8v6j-config-w58gr" Dec 01 08:58:33 crc kubenswrapper[4873]: I1201 08:58:33.845163 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6ae1598f-7693-4112-bc8d-a46e3c5e7f57-var-run-ovn\") pod \"ovn-controller-w8v6j-config-w58gr\" (UID: \"6ae1598f-7693-4112-bc8d-a46e3c5e7f57\") " pod="openstack/ovn-controller-w8v6j-config-w58gr" Dec 01 08:58:33 crc kubenswrapper[4873]: I1201 08:58:33.845186 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6ae1598f-7693-4112-bc8d-a46e3c5e7f57-var-run\") pod \"ovn-controller-w8v6j-config-w58gr\" (UID: \"6ae1598f-7693-4112-bc8d-a46e3c5e7f57\") " pod="openstack/ovn-controller-w8v6j-config-w58gr" Dec 01 08:58:33 crc kubenswrapper[4873]: I1201 08:58:33.845294 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6ae1598f-7693-4112-bc8d-a46e3c5e7f57-var-log-ovn\") pod \"ovn-controller-w8v6j-config-w58gr\" (UID: \"6ae1598f-7693-4112-bc8d-a46e3c5e7f57\") " pod="openstack/ovn-controller-w8v6j-config-w58gr" Dec 01 08:58:33 crc kubenswrapper[4873]: I1201 08:58:33.846251 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6ae1598f-7693-4112-bc8d-a46e3c5e7f57-additional-scripts\") pod \"ovn-controller-w8v6j-config-w58gr\" (UID: \"6ae1598f-7693-4112-bc8d-a46e3c5e7f57\") " pod="openstack/ovn-controller-w8v6j-config-w58gr" Dec 01 08:58:33 crc kubenswrapper[4873]: I1201 08:58:33.847635 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6ae1598f-7693-4112-bc8d-a46e3c5e7f57-scripts\") pod \"ovn-controller-w8v6j-config-w58gr\" (UID: \"6ae1598f-7693-4112-bc8d-a46e3c5e7f57\") " pod="openstack/ovn-controller-w8v6j-config-w58gr" Dec 01 08:58:33 crc kubenswrapper[4873]: I1201 08:58:33.895443 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dw6t\" (UniqueName: \"kubernetes.io/projected/6ae1598f-7693-4112-bc8d-a46e3c5e7f57-kube-api-access-5dw6t\") pod \"ovn-controller-w8v6j-config-w58gr\" (UID: \"6ae1598f-7693-4112-bc8d-a46e3c5e7f57\") " pod="openstack/ovn-controller-w8v6j-config-w58gr" Dec 01 08:58:34 crc kubenswrapper[4873]: I1201 08:58:34.006835 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-w8v6j-config-w58gr" Dec 01 08:58:38 crc kubenswrapper[4873]: I1201 08:58:38.416233 4873 generic.go:334] "Generic (PLEG): container finished" podID="29423166-c350-44f8-97ff-adb8b2b99165" containerID="00ccca11cf2b2f7b96f7e096e15b6f797e1625129d5561faba84ede8dfacb4e9" exitCode=0 Dec 01 08:58:38 crc kubenswrapper[4873]: I1201 08:58:38.417114 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"29423166-c350-44f8-97ff-adb8b2b99165","Type":"ContainerDied","Data":"00ccca11cf2b2f7b96f7e096e15b6f797e1625129d5561faba84ede8dfacb4e9"} Dec 01 08:58:38 crc kubenswrapper[4873]: I1201 08:58:38.540564 4873 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-w8v6j" podUID="b79364c5-00b2-4ab8-b2f0-aff17ed902b4" containerName="ovn-controller" probeResult="failure" output=< Dec 01 08:58:38 crc kubenswrapper[4873]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 01 08:58:38 crc kubenswrapper[4873]: > Dec 01 08:58:42 crc kubenswrapper[4873]: E1201 08:58:42.552654 4873 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api:current-podified" Dec 01 08:58:42 crc kubenswrapper[4873]: E1201 08:58:42.554050 4873 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s5q8q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-6k8d4_openstack(0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 08:58:42 crc kubenswrapper[4873]: E1201 08:58:42.555365 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-6k8d4" podUID="0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8" Dec 01 08:58:42 crc kubenswrapper[4873]: I1201 08:58:42.687578 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-cbzpc" Dec 01 08:58:42 crc kubenswrapper[4873]: I1201 08:58:42.703860 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-552e-account-create-update-9r7p2" Dec 01 08:58:42 crc kubenswrapper[4873]: I1201 08:58:42.718832 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d1118a3-3429-457f-bfcf-40e9f173798e-operator-scripts\") pod \"9d1118a3-3429-457f-bfcf-40e9f173798e\" (UID: \"9d1118a3-3429-457f-bfcf-40e9f173798e\") " Dec 01 08:58:42 crc kubenswrapper[4873]: I1201 08:58:42.718920 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ld7sv\" (UniqueName: \"kubernetes.io/projected/9d1118a3-3429-457f-bfcf-40e9f173798e-kube-api-access-ld7sv\") pod \"9d1118a3-3429-457f-bfcf-40e9f173798e\" (UID: \"9d1118a3-3429-457f-bfcf-40e9f173798e\") " Dec 01 08:58:42 crc kubenswrapper[4873]: I1201 08:58:42.718972 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-68gd9\" (UniqueName: \"kubernetes.io/projected/98acc7b6-efd7-49b8-bcd5-ae1d67848dd7-kube-api-access-68gd9\") pod \"98acc7b6-efd7-49b8-bcd5-ae1d67848dd7\" (UID: \"98acc7b6-efd7-49b8-bcd5-ae1d67848dd7\") " Dec 01 08:58:42 crc kubenswrapper[4873]: I1201 08:58:42.721157 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d1118a3-3429-457f-bfcf-40e9f173798e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9d1118a3-3429-457f-bfcf-40e9f173798e" (UID: "9d1118a3-3429-457f-bfcf-40e9f173798e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:58:42 crc kubenswrapper[4873]: I1201 08:58:42.727179 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d1118a3-3429-457f-bfcf-40e9f173798e-kube-api-access-ld7sv" (OuterVolumeSpecName: "kube-api-access-ld7sv") pod "9d1118a3-3429-457f-bfcf-40e9f173798e" (UID: "9d1118a3-3429-457f-bfcf-40e9f173798e"). InnerVolumeSpecName "kube-api-access-ld7sv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:58:42 crc kubenswrapper[4873]: I1201 08:58:42.742272 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98acc7b6-efd7-49b8-bcd5-ae1d67848dd7-kube-api-access-68gd9" (OuterVolumeSpecName: "kube-api-access-68gd9") pod "98acc7b6-efd7-49b8-bcd5-ae1d67848dd7" (UID: "98acc7b6-efd7-49b8-bcd5-ae1d67848dd7"). InnerVolumeSpecName "kube-api-access-68gd9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:58:42 crc kubenswrapper[4873]: I1201 08:58:42.821894 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98acc7b6-efd7-49b8-bcd5-ae1d67848dd7-operator-scripts\") pod \"98acc7b6-efd7-49b8-bcd5-ae1d67848dd7\" (UID: \"98acc7b6-efd7-49b8-bcd5-ae1d67848dd7\") " Dec 01 08:58:42 crc kubenswrapper[4873]: I1201 08:58:42.822499 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ld7sv\" (UniqueName: \"kubernetes.io/projected/9d1118a3-3429-457f-bfcf-40e9f173798e-kube-api-access-ld7sv\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:42 crc kubenswrapper[4873]: I1201 08:58:42.822515 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-68gd9\" (UniqueName: \"kubernetes.io/projected/98acc7b6-efd7-49b8-bcd5-ae1d67848dd7-kube-api-access-68gd9\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:42 crc kubenswrapper[4873]: I1201 08:58:42.822526 4873 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d1118a3-3429-457f-bfcf-40e9f173798e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:42 crc kubenswrapper[4873]: I1201 08:58:42.823490 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/98acc7b6-efd7-49b8-bcd5-ae1d67848dd7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "98acc7b6-efd7-49b8-bcd5-ae1d67848dd7" (UID: "98acc7b6-efd7-49b8-bcd5-ae1d67848dd7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:58:42 crc kubenswrapper[4873]: I1201 08:58:42.924573 4873 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98acc7b6-efd7-49b8-bcd5-ae1d67848dd7-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:43 crc kubenswrapper[4873]: I1201 08:58:43.097163 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-w8v6j-config-w58gr"] Dec 01 08:58:43 crc kubenswrapper[4873]: I1201 08:58:43.328480 4873 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-w8v6j" podUID="b79364c5-00b2-4ab8-b2f0-aff17ed902b4" containerName="ovn-controller" probeResult="failure" output=< Dec 01 08:58:43 crc kubenswrapper[4873]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 01 08:58:43 crc kubenswrapper[4873]: > Dec 01 08:58:43 crc kubenswrapper[4873]: I1201 08:58:43.502141 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-552e-account-create-update-9r7p2" event={"ID":"98acc7b6-efd7-49b8-bcd5-ae1d67848dd7","Type":"ContainerDied","Data":"a26163b58ee631f42e044c0b6f4b6045df6fb317856cd049caf0a2ca725c0a49"} Dec 01 08:58:43 crc kubenswrapper[4873]: I1201 08:58:43.502194 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a26163b58ee631f42e044c0b6f4b6045df6fb317856cd049caf0a2ca725c0a49" Dec 01 08:58:43 crc kubenswrapper[4873]: I1201 08:58:43.502273 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-552e-account-create-update-9r7p2" Dec 01 08:58:43 crc kubenswrapper[4873]: I1201 08:58:43.509722 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-w8v6j-config-w58gr" event={"ID":"6ae1598f-7693-4112-bc8d-a46e3c5e7f57","Type":"ContainerStarted","Data":"38e5c17e0c90dadd084d4db66895f6c628fa7e38fe797272d17ed7a2e0cdfe88"} Dec 01 08:58:43 crc kubenswrapper[4873]: I1201 08:58:43.509794 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-w8v6j-config-w58gr" event={"ID":"6ae1598f-7693-4112-bc8d-a46e3c5e7f57","Type":"ContainerStarted","Data":"c551f4fd9ae20348ec91cc34073811659de6804005ce43b684d1154a831ea507"} Dec 01 08:58:43 crc kubenswrapper[4873]: I1201 08:58:43.515089 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"29423166-c350-44f8-97ff-adb8b2b99165","Type":"ContainerStarted","Data":"74d71a8b0d6ed6192d61a25abdf207c390202505626ac490b22be2b31b6322a4"} Dec 01 08:58:43 crc kubenswrapper[4873]: I1201 08:58:43.516223 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:58:43 crc kubenswrapper[4873]: I1201 08:58:43.520755 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-cbzpc" Dec 01 08:58:43 crc kubenswrapper[4873]: I1201 08:58:43.524178 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-cbzpc" event={"ID":"9d1118a3-3429-457f-bfcf-40e9f173798e","Type":"ContainerDied","Data":"333c58f68b5094cc4b4efecc77944b615d7eeb3bcc75a0d8084b460616d57439"} Dec 01 08:58:43 crc kubenswrapper[4873]: I1201 08:58:43.524238 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="333c58f68b5094cc4b4efecc77944b615d7eeb3bcc75a0d8084b460616d57439" Dec 01 08:58:43 crc kubenswrapper[4873]: I1201 08:58:43.531123 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-w8v6j-config-w58gr" podStartSLOduration=10.531099725 podStartE2EDuration="10.531099725s" podCreationTimestamp="2025-12-01 08:58:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:58:43.526358493 +0000 UTC m=+1099.428467022" watchObservedRunningTime="2025-12-01 08:58:43.531099725 +0000 UTC m=+1099.433208274" Dec 01 08:58:43 crc kubenswrapper[4873]: E1201 08:58:43.540733 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api:current-podified\\\"\"" pod="openstack/glance-db-sync-6k8d4" podUID="0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8" Dec 01 08:58:43 crc kubenswrapper[4873]: I1201 08:58:43.570077 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=-9223371956.284733 podStartE2EDuration="1m20.570042473s" podCreationTimestamp="2025-12-01 08:57:23 +0000 UTC" firstStartedPulling="2025-12-01 08:57:28.128220431 +0000 UTC m=+1024.030328970" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:58:43.55506069 +0000 UTC m=+1099.457169249" watchObservedRunningTime="2025-12-01 08:58:43.570042473 +0000 UTC m=+1099.472151022" Dec 01 08:58:44 crc kubenswrapper[4873]: I1201 08:58:44.531928 4873 generic.go:334] "Generic (PLEG): container finished" podID="6ae1598f-7693-4112-bc8d-a46e3c5e7f57" containerID="38e5c17e0c90dadd084d4db66895f6c628fa7e38fe797272d17ed7a2e0cdfe88" exitCode=0 Dec 01 08:58:44 crc kubenswrapper[4873]: I1201 08:58:44.532006 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-w8v6j-config-w58gr" event={"ID":"6ae1598f-7693-4112-bc8d-a46e3c5e7f57","Type":"ContainerDied","Data":"38e5c17e0c90dadd084d4db66895f6c628fa7e38fe797272d17ed7a2e0cdfe88"} Dec 01 08:58:45 crc kubenswrapper[4873]: I1201 08:58:45.909692 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-w8v6j-config-w58gr" Dec 01 08:58:46 crc kubenswrapper[4873]: I1201 08:58:46.096384 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6ae1598f-7693-4112-bc8d-a46e3c5e7f57-scripts\") pod \"6ae1598f-7693-4112-bc8d-a46e3c5e7f57\" (UID: \"6ae1598f-7693-4112-bc8d-a46e3c5e7f57\") " Dec 01 08:58:46 crc kubenswrapper[4873]: I1201 08:58:46.096860 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6ae1598f-7693-4112-bc8d-a46e3c5e7f57-additional-scripts\") pod \"6ae1598f-7693-4112-bc8d-a46e3c5e7f57\" (UID: \"6ae1598f-7693-4112-bc8d-a46e3c5e7f57\") " Dec 01 08:58:46 crc kubenswrapper[4873]: I1201 08:58:46.097009 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6ae1598f-7693-4112-bc8d-a46e3c5e7f57-var-run-ovn\") pod \"6ae1598f-7693-4112-bc8d-a46e3c5e7f57\" (UID: \"6ae1598f-7693-4112-bc8d-a46e3c5e7f57\") " Dec 01 08:58:46 crc kubenswrapper[4873]: I1201 08:58:46.097148 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6ae1598f-7693-4112-bc8d-a46e3c5e7f57-var-log-ovn\") pod \"6ae1598f-7693-4112-bc8d-a46e3c5e7f57\" (UID: \"6ae1598f-7693-4112-bc8d-a46e3c5e7f57\") " Dec 01 08:58:46 crc kubenswrapper[4873]: I1201 08:58:46.097263 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5dw6t\" (UniqueName: \"kubernetes.io/projected/6ae1598f-7693-4112-bc8d-a46e3c5e7f57-kube-api-access-5dw6t\") pod \"6ae1598f-7693-4112-bc8d-a46e3c5e7f57\" (UID: \"6ae1598f-7693-4112-bc8d-a46e3c5e7f57\") " Dec 01 08:58:46 crc kubenswrapper[4873]: I1201 08:58:46.097334 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6ae1598f-7693-4112-bc8d-a46e3c5e7f57-var-run\") pod \"6ae1598f-7693-4112-bc8d-a46e3c5e7f57\" (UID: \"6ae1598f-7693-4112-bc8d-a46e3c5e7f57\") " Dec 01 08:58:46 crc kubenswrapper[4873]: I1201 08:58:46.097903 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6ae1598f-7693-4112-bc8d-a46e3c5e7f57-var-run" (OuterVolumeSpecName: "var-run") pod "6ae1598f-7693-4112-bc8d-a46e3c5e7f57" (UID: "6ae1598f-7693-4112-bc8d-a46e3c5e7f57"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:58:46 crc kubenswrapper[4873]: I1201 08:58:46.099292 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ae1598f-7693-4112-bc8d-a46e3c5e7f57-scripts" (OuterVolumeSpecName: "scripts") pod "6ae1598f-7693-4112-bc8d-a46e3c5e7f57" (UID: "6ae1598f-7693-4112-bc8d-a46e3c5e7f57"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:58:46 crc kubenswrapper[4873]: I1201 08:58:46.100203 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ae1598f-7693-4112-bc8d-a46e3c5e7f57-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "6ae1598f-7693-4112-bc8d-a46e3c5e7f57" (UID: "6ae1598f-7693-4112-bc8d-a46e3c5e7f57"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:58:46 crc kubenswrapper[4873]: I1201 08:58:46.100319 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6ae1598f-7693-4112-bc8d-a46e3c5e7f57-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "6ae1598f-7693-4112-bc8d-a46e3c5e7f57" (UID: "6ae1598f-7693-4112-bc8d-a46e3c5e7f57"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:58:46 crc kubenswrapper[4873]: I1201 08:58:46.100404 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6ae1598f-7693-4112-bc8d-a46e3c5e7f57-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "6ae1598f-7693-4112-bc8d-a46e3c5e7f57" (UID: "6ae1598f-7693-4112-bc8d-a46e3c5e7f57"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:58:46 crc kubenswrapper[4873]: I1201 08:58:46.106158 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ae1598f-7693-4112-bc8d-a46e3c5e7f57-kube-api-access-5dw6t" (OuterVolumeSpecName: "kube-api-access-5dw6t") pod "6ae1598f-7693-4112-bc8d-a46e3c5e7f57" (UID: "6ae1598f-7693-4112-bc8d-a46e3c5e7f57"). InnerVolumeSpecName "kube-api-access-5dw6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:58:46 crc kubenswrapper[4873]: I1201 08:58:46.199974 4873 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6ae1598f-7693-4112-bc8d-a46e3c5e7f57-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:46 crc kubenswrapper[4873]: I1201 08:58:46.200295 4873 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6ae1598f-7693-4112-bc8d-a46e3c5e7f57-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:46 crc kubenswrapper[4873]: I1201 08:58:46.200398 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5dw6t\" (UniqueName: \"kubernetes.io/projected/6ae1598f-7693-4112-bc8d-a46e3c5e7f57-kube-api-access-5dw6t\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:46 crc kubenswrapper[4873]: I1201 08:58:46.200486 4873 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6ae1598f-7693-4112-bc8d-a46e3c5e7f57-var-run\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:46 crc kubenswrapper[4873]: I1201 08:58:46.200564 4873 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6ae1598f-7693-4112-bc8d-a46e3c5e7f57-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:46 crc kubenswrapper[4873]: I1201 08:58:46.200641 4873 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6ae1598f-7693-4112-bc8d-a46e3c5e7f57-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:46 crc kubenswrapper[4873]: I1201 08:58:46.212573 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-w8v6j-config-w58gr"] Dec 01 08:58:46 crc kubenswrapper[4873]: I1201 08:58:46.220769 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-w8v6j-config-w58gr"] Dec 01 08:58:46 crc kubenswrapper[4873]: I1201 08:58:46.443898 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ae1598f-7693-4112-bc8d-a46e3c5e7f57" path="/var/lib/kubelet/pods/6ae1598f-7693-4112-bc8d-a46e3c5e7f57/volumes" Dec 01 08:58:46 crc kubenswrapper[4873]: I1201 08:58:46.551325 4873 scope.go:117] "RemoveContainer" containerID="38e5c17e0c90dadd084d4db66895f6c628fa7e38fe797272d17ed7a2e0cdfe88" Dec 01 08:58:46 crc kubenswrapper[4873]: I1201 08:58:46.551518 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-w8v6j-config-w58gr" Dec 01 08:58:48 crc kubenswrapper[4873]: I1201 08:58:48.323157 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-w8v6j" Dec 01 08:58:48 crc kubenswrapper[4873]: I1201 08:58:48.576172 4873 generic.go:334] "Generic (PLEG): container finished" podID="6d27dec2-61e5-44ae-81df-8fee227ce3d8" containerID="cfec876654e058bc3231e41fa25327a07bbec96f72ff60c123f47775b7ad0e34" exitCode=0 Dec 01 08:58:48 crc kubenswrapper[4873]: I1201 08:58:48.576244 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6d27dec2-61e5-44ae-81df-8fee227ce3d8","Type":"ContainerDied","Data":"cfec876654e058bc3231e41fa25327a07bbec96f72ff60c123f47775b7ad0e34"} Dec 01 08:58:49 crc kubenswrapper[4873]: I1201 08:58:49.586833 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6d27dec2-61e5-44ae-81df-8fee227ce3d8","Type":"ContainerStarted","Data":"59527d362e9a669ae5d6a07c74aacb56d24cf16f7963e8390b922ac5ab0c72b8"} Dec 01 08:58:49 crc kubenswrapper[4873]: I1201 08:58:49.587624 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 01 08:58:49 crc kubenswrapper[4873]: I1201 08:58:49.614216 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=49.391057507 podStartE2EDuration="1m27.614189265s" podCreationTimestamp="2025-12-01 08:57:22 +0000 UTC" firstStartedPulling="2025-12-01 08:57:24.841669861 +0000 UTC m=+1020.743778400" lastFinishedPulling="2025-12-01 08:58:03.064801619 +0000 UTC m=+1058.966910158" observedRunningTime="2025-12-01 08:58:49.608640195 +0000 UTC m=+1105.510748744" watchObservedRunningTime="2025-12-01 08:58:49.614189265 +0000 UTC m=+1105.516297804" Dec 01 08:58:54 crc kubenswrapper[4873]: I1201 08:58:54.768291 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:58:57 crc kubenswrapper[4873]: I1201 08:58:57.679830 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-6k8d4" event={"ID":"0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8","Type":"ContainerStarted","Data":"e81edd13885b27ff4960dd15fb62847773d51dce54133f76e250799efff23ff7"} Dec 01 08:58:57 crc kubenswrapper[4873]: I1201 08:58:57.709833 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-6k8d4" podStartSLOduration=1.683411285 podStartE2EDuration="34.709810809s" podCreationTimestamp="2025-12-01 08:58:23 +0000 UTC" firstStartedPulling="2025-12-01 08:58:24.01605908 +0000 UTC m=+1079.918167619" lastFinishedPulling="2025-12-01 08:58:57.042458604 +0000 UTC m=+1112.944567143" observedRunningTime="2025-12-01 08:58:57.7026392 +0000 UTC m=+1113.604747739" watchObservedRunningTime="2025-12-01 08:58:57.709810809 +0000 UTC m=+1113.611919348" Dec 01 08:59:01 crc kubenswrapper[4873]: I1201 08:59:01.058625 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:59:01 crc kubenswrapper[4873]: I1201 08:59:01.059139 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:59:04 crc kubenswrapper[4873]: I1201 08:59:04.288869 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 01 08:59:04 crc kubenswrapper[4873]: I1201 08:59:04.739583 4873 generic.go:334] "Generic (PLEG): container finished" podID="0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8" containerID="e81edd13885b27ff4960dd15fb62847773d51dce54133f76e250799efff23ff7" exitCode=0 Dec 01 08:59:04 crc kubenswrapper[4873]: I1201 08:59:04.739634 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-6k8d4" event={"ID":"0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8","Type":"ContainerDied","Data":"e81edd13885b27ff4960dd15fb62847773d51dce54133f76e250799efff23ff7"} Dec 01 08:59:04 crc kubenswrapper[4873]: I1201 08:59:04.753115 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-m9cfn"] Dec 01 08:59:04 crc kubenswrapper[4873]: E1201 08:59:04.753491 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d1118a3-3429-457f-bfcf-40e9f173798e" containerName="mariadb-database-create" Dec 01 08:59:04 crc kubenswrapper[4873]: I1201 08:59:04.753505 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d1118a3-3429-457f-bfcf-40e9f173798e" containerName="mariadb-database-create" Dec 01 08:59:04 crc kubenswrapper[4873]: E1201 08:59:04.753534 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ae1598f-7693-4112-bc8d-a46e3c5e7f57" containerName="ovn-config" Dec 01 08:59:04 crc kubenswrapper[4873]: I1201 08:59:04.753542 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ae1598f-7693-4112-bc8d-a46e3c5e7f57" containerName="ovn-config" Dec 01 08:59:04 crc kubenswrapper[4873]: E1201 08:59:04.753556 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98acc7b6-efd7-49b8-bcd5-ae1d67848dd7" containerName="mariadb-account-create-update" Dec 01 08:59:04 crc kubenswrapper[4873]: I1201 08:59:04.753566 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="98acc7b6-efd7-49b8-bcd5-ae1d67848dd7" containerName="mariadb-account-create-update" Dec 01 08:59:04 crc kubenswrapper[4873]: I1201 08:59:04.753721 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="98acc7b6-efd7-49b8-bcd5-ae1d67848dd7" containerName="mariadb-account-create-update" Dec 01 08:59:04 crc kubenswrapper[4873]: I1201 08:59:04.753741 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ae1598f-7693-4112-bc8d-a46e3c5e7f57" containerName="ovn-config" Dec 01 08:59:04 crc kubenswrapper[4873]: I1201 08:59:04.753757 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d1118a3-3429-457f-bfcf-40e9f173798e" containerName="mariadb-database-create" Dec 01 08:59:04 crc kubenswrapper[4873]: I1201 08:59:04.754333 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-m9cfn" Dec 01 08:59:04 crc kubenswrapper[4873]: I1201 08:59:04.772691 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-m9cfn"] Dec 01 08:59:04 crc kubenswrapper[4873]: I1201 08:59:04.787581 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-108c-account-create-update-cmrw2"] Dec 01 08:59:04 crc kubenswrapper[4873]: I1201 08:59:04.789365 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-108c-account-create-update-cmrw2" Dec 01 08:59:04 crc kubenswrapper[4873]: I1201 08:59:04.794072 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 01 08:59:04 crc kubenswrapper[4873]: I1201 08:59:04.803911 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a20240ed-c3f5-444d-854d-718d0f5cbace-operator-scripts\") pod \"barbican-db-create-m9cfn\" (UID: \"a20240ed-c3f5-444d-854d-718d0f5cbace\") " pod="openstack/barbican-db-create-m9cfn" Dec 01 08:59:04 crc kubenswrapper[4873]: I1201 08:59:04.803980 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ph6td\" (UniqueName: \"kubernetes.io/projected/a20240ed-c3f5-444d-854d-718d0f5cbace-kube-api-access-ph6td\") pod \"barbican-db-create-m9cfn\" (UID: \"a20240ed-c3f5-444d-854d-718d0f5cbace\") " pod="openstack/barbican-db-create-m9cfn" Dec 01 08:59:04 crc kubenswrapper[4873]: I1201 08:59:04.813329 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-108c-account-create-update-cmrw2"] Dec 01 08:59:04 crc kubenswrapper[4873]: I1201 08:59:04.848141 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-h9jj9"] Dec 01 08:59:04 crc kubenswrapper[4873]: I1201 08:59:04.867402 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-h9jj9" Dec 01 08:59:04 crc kubenswrapper[4873]: I1201 08:59:04.916561 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58mkh\" (UniqueName: \"kubernetes.io/projected/b5650152-030e-42d2-8db1-b80bfa18bc6f-kube-api-access-58mkh\") pod \"barbican-108c-account-create-update-cmrw2\" (UID: \"b5650152-030e-42d2-8db1-b80bfa18bc6f\") " pod="openstack/barbican-108c-account-create-update-cmrw2" Dec 01 08:59:04 crc kubenswrapper[4873]: I1201 08:59:04.916723 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/860ae34b-0993-4a11-ac3a-3774804a0860-operator-scripts\") pod \"cinder-db-create-h9jj9\" (UID: \"860ae34b-0993-4a11-ac3a-3774804a0860\") " pod="openstack/cinder-db-create-h9jj9" Dec 01 08:59:04 crc kubenswrapper[4873]: I1201 08:59:04.922123 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a20240ed-c3f5-444d-854d-718d0f5cbace-operator-scripts\") pod \"barbican-db-create-m9cfn\" (UID: \"a20240ed-c3f5-444d-854d-718d0f5cbace\") " pod="openstack/barbican-db-create-m9cfn" Dec 01 08:59:04 crc kubenswrapper[4873]: I1201 08:59:04.926553 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a20240ed-c3f5-444d-854d-718d0f5cbace-operator-scripts\") pod \"barbican-db-create-m9cfn\" (UID: \"a20240ed-c3f5-444d-854d-718d0f5cbace\") " pod="openstack/barbican-db-create-m9cfn" Dec 01 08:59:04 crc kubenswrapper[4873]: I1201 08:59:04.926711 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ph6td\" (UniqueName: \"kubernetes.io/projected/a20240ed-c3f5-444d-854d-718d0f5cbace-kube-api-access-ph6td\") pod \"barbican-db-create-m9cfn\" (UID: \"a20240ed-c3f5-444d-854d-718d0f5cbace\") " pod="openstack/barbican-db-create-m9cfn" Dec 01 08:59:04 crc kubenswrapper[4873]: I1201 08:59:04.926860 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-np66w\" (UniqueName: \"kubernetes.io/projected/860ae34b-0993-4a11-ac3a-3774804a0860-kube-api-access-np66w\") pod \"cinder-db-create-h9jj9\" (UID: \"860ae34b-0993-4a11-ac3a-3774804a0860\") " pod="openstack/cinder-db-create-h9jj9" Dec 01 08:59:04 crc kubenswrapper[4873]: I1201 08:59:04.926967 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5650152-030e-42d2-8db1-b80bfa18bc6f-operator-scripts\") pod \"barbican-108c-account-create-update-cmrw2\" (UID: \"b5650152-030e-42d2-8db1-b80bfa18bc6f\") " pod="openstack/barbican-108c-account-create-update-cmrw2" Dec 01 08:59:04 crc kubenswrapper[4873]: I1201 08:59:04.955186 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-h9jj9"] Dec 01 08:59:04 crc kubenswrapper[4873]: I1201 08:59:04.974181 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ph6td\" (UniqueName: \"kubernetes.io/projected/a20240ed-c3f5-444d-854d-718d0f5cbace-kube-api-access-ph6td\") pod \"barbican-db-create-m9cfn\" (UID: \"a20240ed-c3f5-444d-854d-718d0f5cbace\") " pod="openstack/barbican-db-create-m9cfn" Dec 01 08:59:04 crc kubenswrapper[4873]: I1201 08:59:04.988093 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-eb5b-account-create-update-q6hf2"] Dec 01 08:59:04 crc kubenswrapper[4873]: I1201 08:59:04.989615 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-eb5b-account-create-update-q6hf2" Dec 01 08:59:04 crc kubenswrapper[4873]: I1201 08:59:04.996058 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.018119 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-eb5b-account-create-update-q6hf2"] Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.029327 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/860ae34b-0993-4a11-ac3a-3774804a0860-operator-scripts\") pod \"cinder-db-create-h9jj9\" (UID: \"860ae34b-0993-4a11-ac3a-3774804a0860\") " pod="openstack/cinder-db-create-h9jj9" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.029419 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-np66w\" (UniqueName: \"kubernetes.io/projected/860ae34b-0993-4a11-ac3a-3774804a0860-kube-api-access-np66w\") pod \"cinder-db-create-h9jj9\" (UID: \"860ae34b-0993-4a11-ac3a-3774804a0860\") " pod="openstack/cinder-db-create-h9jj9" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.029468 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/17c0c1a0-6746-4e83-aa1e-419bc703ac35-operator-scripts\") pod \"cinder-eb5b-account-create-update-q6hf2\" (UID: \"17c0c1a0-6746-4e83-aa1e-419bc703ac35\") " pod="openstack/cinder-eb5b-account-create-update-q6hf2" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.029509 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5650152-030e-42d2-8db1-b80bfa18bc6f-operator-scripts\") pod \"barbican-108c-account-create-update-cmrw2\" (UID: \"b5650152-030e-42d2-8db1-b80bfa18bc6f\") " pod="openstack/barbican-108c-account-create-update-cmrw2" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.029537 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l52rq\" (UniqueName: \"kubernetes.io/projected/17c0c1a0-6746-4e83-aa1e-419bc703ac35-kube-api-access-l52rq\") pod \"cinder-eb5b-account-create-update-q6hf2\" (UID: \"17c0c1a0-6746-4e83-aa1e-419bc703ac35\") " pod="openstack/cinder-eb5b-account-create-update-q6hf2" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.029586 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58mkh\" (UniqueName: \"kubernetes.io/projected/b5650152-030e-42d2-8db1-b80bfa18bc6f-kube-api-access-58mkh\") pod \"barbican-108c-account-create-update-cmrw2\" (UID: \"b5650152-030e-42d2-8db1-b80bfa18bc6f\") " pod="openstack/barbican-108c-account-create-update-cmrw2" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.030712 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/860ae34b-0993-4a11-ac3a-3774804a0860-operator-scripts\") pod \"cinder-db-create-h9jj9\" (UID: \"860ae34b-0993-4a11-ac3a-3774804a0860\") " pod="openstack/cinder-db-create-h9jj9" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.031290 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5650152-030e-42d2-8db1-b80bfa18bc6f-operator-scripts\") pod \"barbican-108c-account-create-update-cmrw2\" (UID: \"b5650152-030e-42d2-8db1-b80bfa18bc6f\") " pod="openstack/barbican-108c-account-create-update-cmrw2" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.053572 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-np66w\" (UniqueName: \"kubernetes.io/projected/860ae34b-0993-4a11-ac3a-3774804a0860-kube-api-access-np66w\") pod \"cinder-db-create-h9jj9\" (UID: \"860ae34b-0993-4a11-ac3a-3774804a0860\") " pod="openstack/cinder-db-create-h9jj9" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.056660 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-wshpl"] Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.057708 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58mkh\" (UniqueName: \"kubernetes.io/projected/b5650152-030e-42d2-8db1-b80bfa18bc6f-kube-api-access-58mkh\") pod \"barbican-108c-account-create-update-cmrw2\" (UID: \"b5650152-030e-42d2-8db1-b80bfa18bc6f\") " pod="openstack/barbican-108c-account-create-update-cmrw2" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.057891 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-wshpl" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.065522 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.065957 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.066119 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.066236 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-9jwfx" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.072621 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-wshpl"] Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.080443 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-m9cfn" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.109770 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-108c-account-create-update-cmrw2" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.133589 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kd4g\" (UniqueName: \"kubernetes.io/projected/1b86ab79-9562-43e9-a624-f0bc8e1e16c3-kube-api-access-2kd4g\") pod \"keystone-db-sync-wshpl\" (UID: \"1b86ab79-9562-43e9-a624-f0bc8e1e16c3\") " pod="openstack/keystone-db-sync-wshpl" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.133706 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b86ab79-9562-43e9-a624-f0bc8e1e16c3-config-data\") pod \"keystone-db-sync-wshpl\" (UID: \"1b86ab79-9562-43e9-a624-f0bc8e1e16c3\") " pod="openstack/keystone-db-sync-wshpl" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.133766 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/17c0c1a0-6746-4e83-aa1e-419bc703ac35-operator-scripts\") pod \"cinder-eb5b-account-create-update-q6hf2\" (UID: \"17c0c1a0-6746-4e83-aa1e-419bc703ac35\") " pod="openstack/cinder-eb5b-account-create-update-q6hf2" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.133807 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l52rq\" (UniqueName: \"kubernetes.io/projected/17c0c1a0-6746-4e83-aa1e-419bc703ac35-kube-api-access-l52rq\") pod \"cinder-eb5b-account-create-update-q6hf2\" (UID: \"17c0c1a0-6746-4e83-aa1e-419bc703ac35\") " pod="openstack/cinder-eb5b-account-create-update-q6hf2" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.133840 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b86ab79-9562-43e9-a624-f0bc8e1e16c3-combined-ca-bundle\") pod \"keystone-db-sync-wshpl\" (UID: \"1b86ab79-9562-43e9-a624-f0bc8e1e16c3\") " pod="openstack/keystone-db-sync-wshpl" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.134832 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/17c0c1a0-6746-4e83-aa1e-419bc703ac35-operator-scripts\") pod \"cinder-eb5b-account-create-update-q6hf2\" (UID: \"17c0c1a0-6746-4e83-aa1e-419bc703ac35\") " pod="openstack/cinder-eb5b-account-create-update-q6hf2" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.164432 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-q6bwr"] Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.165894 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-q6bwr" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.165968 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l52rq\" (UniqueName: \"kubernetes.io/projected/17c0c1a0-6746-4e83-aa1e-419bc703ac35-kube-api-access-l52rq\") pod \"cinder-eb5b-account-create-update-q6hf2\" (UID: \"17c0c1a0-6746-4e83-aa1e-419bc703ac35\") " pod="openstack/cinder-eb5b-account-create-update-q6hf2" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.193271 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-3c32-account-create-update-vl8cf"] Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.195244 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-3c32-account-create-update-vl8cf" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.198256 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.203800 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-q6bwr"] Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.214846 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-3c32-account-create-update-vl8cf"] Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.228462 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-h9jj9" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.235480 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5f95\" (UniqueName: \"kubernetes.io/projected/69e4256b-7143-4d91-91a0-b4ffdeb36511-kube-api-access-f5f95\") pod \"neutron-3c32-account-create-update-vl8cf\" (UID: \"69e4256b-7143-4d91-91a0-b4ffdeb36511\") " pod="openstack/neutron-3c32-account-create-update-vl8cf" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.235553 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b86ab79-9562-43e9-a624-f0bc8e1e16c3-config-data\") pod \"keystone-db-sync-wshpl\" (UID: \"1b86ab79-9562-43e9-a624-f0bc8e1e16c3\") " pod="openstack/keystone-db-sync-wshpl" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.235580 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/69e4256b-7143-4d91-91a0-b4ffdeb36511-operator-scripts\") pod \"neutron-3c32-account-create-update-vl8cf\" (UID: \"69e4256b-7143-4d91-91a0-b4ffdeb36511\") " pod="openstack/neutron-3c32-account-create-update-vl8cf" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.235662 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvxpc\" (UniqueName: \"kubernetes.io/projected/570a0599-aafa-47fe-8b7a-26056d6a1723-kube-api-access-pvxpc\") pod \"neutron-db-create-q6bwr\" (UID: \"570a0599-aafa-47fe-8b7a-26056d6a1723\") " pod="openstack/neutron-db-create-q6bwr" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.235698 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b86ab79-9562-43e9-a624-f0bc8e1e16c3-combined-ca-bundle\") pod \"keystone-db-sync-wshpl\" (UID: \"1b86ab79-9562-43e9-a624-f0bc8e1e16c3\") " pod="openstack/keystone-db-sync-wshpl" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.235732 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kd4g\" (UniqueName: \"kubernetes.io/projected/1b86ab79-9562-43e9-a624-f0bc8e1e16c3-kube-api-access-2kd4g\") pod \"keystone-db-sync-wshpl\" (UID: \"1b86ab79-9562-43e9-a624-f0bc8e1e16c3\") " pod="openstack/keystone-db-sync-wshpl" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.235759 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/570a0599-aafa-47fe-8b7a-26056d6a1723-operator-scripts\") pod \"neutron-db-create-q6bwr\" (UID: \"570a0599-aafa-47fe-8b7a-26056d6a1723\") " pod="openstack/neutron-db-create-q6bwr" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.240477 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b86ab79-9562-43e9-a624-f0bc8e1e16c3-combined-ca-bundle\") pod \"keystone-db-sync-wshpl\" (UID: \"1b86ab79-9562-43e9-a624-f0bc8e1e16c3\") " pod="openstack/keystone-db-sync-wshpl" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.240616 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b86ab79-9562-43e9-a624-f0bc8e1e16c3-config-data\") pod \"keystone-db-sync-wshpl\" (UID: \"1b86ab79-9562-43e9-a624-f0bc8e1e16c3\") " pod="openstack/keystone-db-sync-wshpl" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.295400 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kd4g\" (UniqueName: \"kubernetes.io/projected/1b86ab79-9562-43e9-a624-f0bc8e1e16c3-kube-api-access-2kd4g\") pod \"keystone-db-sync-wshpl\" (UID: \"1b86ab79-9562-43e9-a624-f0bc8e1e16c3\") " pod="openstack/keystone-db-sync-wshpl" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.309135 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-eb5b-account-create-update-q6hf2" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.337009 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvxpc\" (UniqueName: \"kubernetes.io/projected/570a0599-aafa-47fe-8b7a-26056d6a1723-kube-api-access-pvxpc\") pod \"neutron-db-create-q6bwr\" (UID: \"570a0599-aafa-47fe-8b7a-26056d6a1723\") " pod="openstack/neutron-db-create-q6bwr" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.337140 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/570a0599-aafa-47fe-8b7a-26056d6a1723-operator-scripts\") pod \"neutron-db-create-q6bwr\" (UID: \"570a0599-aafa-47fe-8b7a-26056d6a1723\") " pod="openstack/neutron-db-create-q6bwr" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.337193 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5f95\" (UniqueName: \"kubernetes.io/projected/69e4256b-7143-4d91-91a0-b4ffdeb36511-kube-api-access-f5f95\") pod \"neutron-3c32-account-create-update-vl8cf\" (UID: \"69e4256b-7143-4d91-91a0-b4ffdeb36511\") " pod="openstack/neutron-3c32-account-create-update-vl8cf" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.337223 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/69e4256b-7143-4d91-91a0-b4ffdeb36511-operator-scripts\") pod \"neutron-3c32-account-create-update-vl8cf\" (UID: \"69e4256b-7143-4d91-91a0-b4ffdeb36511\") " pod="openstack/neutron-3c32-account-create-update-vl8cf" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.338081 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/69e4256b-7143-4d91-91a0-b4ffdeb36511-operator-scripts\") pod \"neutron-3c32-account-create-update-vl8cf\" (UID: \"69e4256b-7143-4d91-91a0-b4ffdeb36511\") " pod="openstack/neutron-3c32-account-create-update-vl8cf" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.341296 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/570a0599-aafa-47fe-8b7a-26056d6a1723-operator-scripts\") pod \"neutron-db-create-q6bwr\" (UID: \"570a0599-aafa-47fe-8b7a-26056d6a1723\") " pod="openstack/neutron-db-create-q6bwr" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.360544 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvxpc\" (UniqueName: \"kubernetes.io/projected/570a0599-aafa-47fe-8b7a-26056d6a1723-kube-api-access-pvxpc\") pod \"neutron-db-create-q6bwr\" (UID: \"570a0599-aafa-47fe-8b7a-26056d6a1723\") " pod="openstack/neutron-db-create-q6bwr" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.364832 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5f95\" (UniqueName: \"kubernetes.io/projected/69e4256b-7143-4d91-91a0-b4ffdeb36511-kube-api-access-f5f95\") pod \"neutron-3c32-account-create-update-vl8cf\" (UID: \"69e4256b-7143-4d91-91a0-b4ffdeb36511\") " pod="openstack/neutron-3c32-account-create-update-vl8cf" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.540006 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-wshpl" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.608480 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-q6bwr" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.619179 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-3c32-account-create-update-vl8cf" Dec 01 08:59:05 crc kubenswrapper[4873]: I1201 08:59:05.928520 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-108c-account-create-update-cmrw2"] Dec 01 08:59:06 crc kubenswrapper[4873]: I1201 08:59:06.001876 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-m9cfn"] Dec 01 08:59:06 crc kubenswrapper[4873]: I1201 08:59:06.050289 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-h9jj9"] Dec 01 08:59:06 crc kubenswrapper[4873]: I1201 08:59:06.066852 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-eb5b-account-create-update-q6hf2"] Dec 01 08:59:06 crc kubenswrapper[4873]: I1201 08:59:06.503225 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-wshpl"] Dec 01 08:59:06 crc kubenswrapper[4873]: I1201 08:59:06.604456 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-6k8d4" Dec 01 08:59:06 crc kubenswrapper[4873]: I1201 08:59:06.622795 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s5q8q\" (UniqueName: \"kubernetes.io/projected/0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8-kube-api-access-s5q8q\") pod \"0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8\" (UID: \"0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8\") " Dec 01 08:59:06 crc kubenswrapper[4873]: I1201 08:59:06.622862 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8-db-sync-config-data\") pod \"0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8\" (UID: \"0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8\") " Dec 01 08:59:06 crc kubenswrapper[4873]: I1201 08:59:06.622928 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8-combined-ca-bundle\") pod \"0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8\" (UID: \"0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8\") " Dec 01 08:59:06 crc kubenswrapper[4873]: I1201 08:59:06.623056 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8-config-data\") pod \"0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8\" (UID: \"0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8\") " Dec 01 08:59:06 crc kubenswrapper[4873]: I1201 08:59:06.633081 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8" (UID: "0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:59:06 crc kubenswrapper[4873]: I1201 08:59:06.637511 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-3c32-account-create-update-vl8cf"] Dec 01 08:59:06 crc kubenswrapper[4873]: I1201 08:59:06.640811 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8-kube-api-access-s5q8q" (OuterVolumeSpecName: "kube-api-access-s5q8q") pod "0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8" (UID: "0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8"). InnerVolumeSpecName "kube-api-access-s5q8q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:59:06 crc kubenswrapper[4873]: I1201 08:59:06.699086 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-q6bwr"] Dec 01 08:59:06 crc kubenswrapper[4873]: I1201 08:59:06.745411 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s5q8q\" (UniqueName: \"kubernetes.io/projected/0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8-kube-api-access-s5q8q\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:06 crc kubenswrapper[4873]: I1201 08:59:06.745460 4873 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:06 crc kubenswrapper[4873]: I1201 08:59:06.747308 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8" (UID: "0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:59:06 crc kubenswrapper[4873]: I1201 08:59:06.797802 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8-config-data" (OuterVolumeSpecName: "config-data") pod "0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8" (UID: "0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:59:06 crc kubenswrapper[4873]: I1201 08:59:06.863328 4873 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:06 crc kubenswrapper[4873]: I1201 08:59:06.863376 4873 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:07 crc kubenswrapper[4873]: I1201 08:59:07.056691 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-108c-account-create-update-cmrw2" event={"ID":"b5650152-030e-42d2-8db1-b80bfa18bc6f","Type":"ContainerStarted","Data":"67014367ee4a3300e046b29a2dbfb826636377bb9001a425a2675910de3226fa"} Dec 01 08:59:07 crc kubenswrapper[4873]: I1201 08:59:07.056758 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-108c-account-create-update-cmrw2" event={"ID":"b5650152-030e-42d2-8db1-b80bfa18bc6f","Type":"ContainerStarted","Data":"64d3f102ff960b2291f754d88e27ed61be361757a44096601615129b751f88f8"} Dec 01 08:59:07 crc kubenswrapper[4873]: I1201 08:59:07.060664 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-3c32-account-create-update-vl8cf" event={"ID":"69e4256b-7143-4d91-91a0-b4ffdeb36511","Type":"ContainerStarted","Data":"acaca650ecfe205574dc9441965a62e774e2bfd43e65ef530b86d70dcfb00a6c"} Dec 01 08:59:07 crc kubenswrapper[4873]: I1201 08:59:07.091799 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-eb5b-account-create-update-q6hf2" event={"ID":"17c0c1a0-6746-4e83-aa1e-419bc703ac35","Type":"ContainerStarted","Data":"d3556f9f6736ecd7a85bdac799730eeae69062e0c3e744ccf474db1e4a374bbd"} Dec 01 08:59:07 crc kubenswrapper[4873]: I1201 08:59:07.091875 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-eb5b-account-create-update-q6hf2" event={"ID":"17c0c1a0-6746-4e83-aa1e-419bc703ac35","Type":"ContainerStarted","Data":"9566b575c98cb13799c4b3c5b7371a10491dd128e237a17fb45b37db5beea866"} Dec 01 08:59:07 crc kubenswrapper[4873]: I1201 08:59:07.099337 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-108c-account-create-update-cmrw2" podStartSLOduration=3.096991726 podStartE2EDuration="3.096991726s" podCreationTimestamp="2025-12-01 08:59:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:59:07.090295848 +0000 UTC m=+1122.992404387" watchObservedRunningTime="2025-12-01 08:59:07.096991726 +0000 UTC m=+1122.999100265" Dec 01 08:59:07 crc kubenswrapper[4873]: I1201 08:59:07.114182 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-q6bwr" event={"ID":"570a0599-aafa-47fe-8b7a-26056d6a1723","Type":"ContainerStarted","Data":"37671b622cdb11993e87bc46503d759901179105f757d67329e96cc990a8c8d9"} Dec 01 08:59:07 crc kubenswrapper[4873]: I1201 08:59:07.146871 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-m9cfn" event={"ID":"a20240ed-c3f5-444d-854d-718d0f5cbace","Type":"ContainerStarted","Data":"243d39e2e5e4ccb73cb1c86fc2ff0271389dcbd153fc38dc44f58b96c2a27060"} Dec 01 08:59:07 crc kubenswrapper[4873]: I1201 08:59:07.146942 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-m9cfn" event={"ID":"a20240ed-c3f5-444d-854d-718d0f5cbace","Type":"ContainerStarted","Data":"ea38facb65915324d92cd0d34edfbca41a5ae04dc7a0e50511be5da3dd4af76e"} Dec 01 08:59:07 crc kubenswrapper[4873]: I1201 08:59:07.157006 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-eb5b-account-create-update-q6hf2" podStartSLOduration=3.15697927 podStartE2EDuration="3.15697927s" podCreationTimestamp="2025-12-01 08:59:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:59:07.133300682 +0000 UTC m=+1123.035409221" watchObservedRunningTime="2025-12-01 08:59:07.15697927 +0000 UTC m=+1123.059087809" Dec 01 08:59:07 crc kubenswrapper[4873]: I1201 08:59:07.172029 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-h9jj9" event={"ID":"860ae34b-0993-4a11-ac3a-3774804a0860","Type":"ContainerStarted","Data":"08894b35673c62becdf6e89e1d7ceaba3675395a718195b458b0423be85453ac"} Dec 01 08:59:07 crc kubenswrapper[4873]: I1201 08:59:07.172091 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-h9jj9" event={"ID":"860ae34b-0993-4a11-ac3a-3774804a0860","Type":"ContainerStarted","Data":"ecd836ffe8b4facb66082b9f3a97c9bec85993c050228026c2d30fe48f30eb9c"} Dec 01 08:59:07 crc kubenswrapper[4873]: I1201 08:59:07.174372 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-m9cfn" podStartSLOduration=3.17435189 podStartE2EDuration="3.17435189s" podCreationTimestamp="2025-12-01 08:59:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:59:07.170180341 +0000 UTC m=+1123.072288880" watchObservedRunningTime="2025-12-01 08:59:07.17435189 +0000 UTC m=+1123.076460429" Dec 01 08:59:07 crc kubenswrapper[4873]: I1201 08:59:07.178370 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-6k8d4" event={"ID":"0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8","Type":"ContainerDied","Data":"d6aaf3fd77cfca960bc6f24e53269da0172d1489fa9b03a5ba2c73d244f3bff1"} Dec 01 08:59:07 crc kubenswrapper[4873]: I1201 08:59:07.178432 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d6aaf3fd77cfca960bc6f24e53269da0172d1489fa9b03a5ba2c73d244f3bff1" Dec 01 08:59:07 crc kubenswrapper[4873]: I1201 08:59:07.178554 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-6k8d4" Dec 01 08:59:07 crc kubenswrapper[4873]: I1201 08:59:07.186317 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-wshpl" event={"ID":"1b86ab79-9562-43e9-a624-f0bc8e1e16c3","Type":"ContainerStarted","Data":"3ba2708931673d7bd4dbc73d5c5d4359bceb072ebf22296660d7cad677f79dcb"} Dec 01 08:59:07 crc kubenswrapper[4873]: I1201 08:59:07.218928 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-h9jj9" podStartSLOduration=3.21890037 podStartE2EDuration="3.21890037s" podCreationTimestamp="2025-12-01 08:59:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:59:07.206499718 +0000 UTC m=+1123.108608257" watchObservedRunningTime="2025-12-01 08:59:07.21890037 +0000 UTC m=+1123.121008909" Dec 01 08:59:07 crc kubenswrapper[4873]: I1201 08:59:07.466757 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-54f9b7b8d9-t4rtz"] Dec 01 08:59:07 crc kubenswrapper[4873]: E1201 08:59:07.468451 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8" containerName="glance-db-sync" Dec 01 08:59:07 crc kubenswrapper[4873]: I1201 08:59:07.468477 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8" containerName="glance-db-sync" Dec 01 08:59:07 crc kubenswrapper[4873]: I1201 08:59:07.468686 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8" containerName="glance-db-sync" Dec 01 08:59:07 crc kubenswrapper[4873]: I1201 08:59:07.474691 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54f9b7b8d9-t4rtz" Dec 01 08:59:07 crc kubenswrapper[4873]: I1201 08:59:07.496757 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54f9b7b8d9-t4rtz"] Dec 01 08:59:07 crc kubenswrapper[4873]: I1201 08:59:07.613880 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3c1fee8d-dc8f-4167-8d20-21ff76dd32de-dns-svc\") pod \"dnsmasq-dns-54f9b7b8d9-t4rtz\" (UID: \"3c1fee8d-dc8f-4167-8d20-21ff76dd32de\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-t4rtz" Dec 01 08:59:07 crc kubenswrapper[4873]: I1201 08:59:07.613943 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkz7k\" (UniqueName: \"kubernetes.io/projected/3c1fee8d-dc8f-4167-8d20-21ff76dd32de-kube-api-access-bkz7k\") pod \"dnsmasq-dns-54f9b7b8d9-t4rtz\" (UID: \"3c1fee8d-dc8f-4167-8d20-21ff76dd32de\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-t4rtz" Dec 01 08:59:07 crc kubenswrapper[4873]: I1201 08:59:07.614180 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c1fee8d-dc8f-4167-8d20-21ff76dd32de-config\") pod \"dnsmasq-dns-54f9b7b8d9-t4rtz\" (UID: \"3c1fee8d-dc8f-4167-8d20-21ff76dd32de\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-t4rtz" Dec 01 08:59:07 crc kubenswrapper[4873]: I1201 08:59:07.614292 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3c1fee8d-dc8f-4167-8d20-21ff76dd32de-ovsdbserver-nb\") pod \"dnsmasq-dns-54f9b7b8d9-t4rtz\" (UID: \"3c1fee8d-dc8f-4167-8d20-21ff76dd32de\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-t4rtz" Dec 01 08:59:07 crc kubenswrapper[4873]: I1201 08:59:07.617180 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3c1fee8d-dc8f-4167-8d20-21ff76dd32de-ovsdbserver-sb\") pod \"dnsmasq-dns-54f9b7b8d9-t4rtz\" (UID: \"3c1fee8d-dc8f-4167-8d20-21ff76dd32de\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-t4rtz" Dec 01 08:59:07 crc kubenswrapper[4873]: I1201 08:59:07.721438 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3c1fee8d-dc8f-4167-8d20-21ff76dd32de-ovsdbserver-nb\") pod \"dnsmasq-dns-54f9b7b8d9-t4rtz\" (UID: \"3c1fee8d-dc8f-4167-8d20-21ff76dd32de\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-t4rtz" Dec 01 08:59:07 crc kubenswrapper[4873]: I1201 08:59:07.721534 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3c1fee8d-dc8f-4167-8d20-21ff76dd32de-ovsdbserver-sb\") pod \"dnsmasq-dns-54f9b7b8d9-t4rtz\" (UID: \"3c1fee8d-dc8f-4167-8d20-21ff76dd32de\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-t4rtz" Dec 01 08:59:07 crc kubenswrapper[4873]: I1201 08:59:07.721578 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3c1fee8d-dc8f-4167-8d20-21ff76dd32de-dns-svc\") pod \"dnsmasq-dns-54f9b7b8d9-t4rtz\" (UID: \"3c1fee8d-dc8f-4167-8d20-21ff76dd32de\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-t4rtz" Dec 01 08:59:07 crc kubenswrapper[4873]: I1201 08:59:07.721607 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkz7k\" (UniqueName: \"kubernetes.io/projected/3c1fee8d-dc8f-4167-8d20-21ff76dd32de-kube-api-access-bkz7k\") pod \"dnsmasq-dns-54f9b7b8d9-t4rtz\" (UID: \"3c1fee8d-dc8f-4167-8d20-21ff76dd32de\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-t4rtz" Dec 01 08:59:07 crc kubenswrapper[4873]: I1201 08:59:07.721634 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c1fee8d-dc8f-4167-8d20-21ff76dd32de-config\") pod \"dnsmasq-dns-54f9b7b8d9-t4rtz\" (UID: \"3c1fee8d-dc8f-4167-8d20-21ff76dd32de\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-t4rtz" Dec 01 08:59:07 crc kubenswrapper[4873]: I1201 08:59:07.722604 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3c1fee8d-dc8f-4167-8d20-21ff76dd32de-ovsdbserver-nb\") pod \"dnsmasq-dns-54f9b7b8d9-t4rtz\" (UID: \"3c1fee8d-dc8f-4167-8d20-21ff76dd32de\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-t4rtz" Dec 01 08:59:07 crc kubenswrapper[4873]: I1201 08:59:07.722623 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c1fee8d-dc8f-4167-8d20-21ff76dd32de-config\") pod \"dnsmasq-dns-54f9b7b8d9-t4rtz\" (UID: \"3c1fee8d-dc8f-4167-8d20-21ff76dd32de\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-t4rtz" Dec 01 08:59:07 crc kubenswrapper[4873]: I1201 08:59:07.722752 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3c1fee8d-dc8f-4167-8d20-21ff76dd32de-ovsdbserver-sb\") pod \"dnsmasq-dns-54f9b7b8d9-t4rtz\" (UID: \"3c1fee8d-dc8f-4167-8d20-21ff76dd32de\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-t4rtz" Dec 01 08:59:07 crc kubenswrapper[4873]: I1201 08:59:07.723191 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3c1fee8d-dc8f-4167-8d20-21ff76dd32de-dns-svc\") pod \"dnsmasq-dns-54f9b7b8d9-t4rtz\" (UID: \"3c1fee8d-dc8f-4167-8d20-21ff76dd32de\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-t4rtz" Dec 01 08:59:07 crc kubenswrapper[4873]: I1201 08:59:07.745059 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkz7k\" (UniqueName: \"kubernetes.io/projected/3c1fee8d-dc8f-4167-8d20-21ff76dd32de-kube-api-access-bkz7k\") pod \"dnsmasq-dns-54f9b7b8d9-t4rtz\" (UID: \"3c1fee8d-dc8f-4167-8d20-21ff76dd32de\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-t4rtz" Dec 01 08:59:07 crc kubenswrapper[4873]: I1201 08:59:07.824941 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54f9b7b8d9-t4rtz" Dec 01 08:59:08 crc kubenswrapper[4873]: I1201 08:59:08.200221 4873 generic.go:334] "Generic (PLEG): container finished" podID="17c0c1a0-6746-4e83-aa1e-419bc703ac35" containerID="d3556f9f6736ecd7a85bdac799730eeae69062e0c3e744ccf474db1e4a374bbd" exitCode=0 Dec 01 08:59:08 crc kubenswrapper[4873]: I1201 08:59:08.200424 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-eb5b-account-create-update-q6hf2" event={"ID":"17c0c1a0-6746-4e83-aa1e-419bc703ac35","Type":"ContainerDied","Data":"d3556f9f6736ecd7a85bdac799730eeae69062e0c3e744ccf474db1e4a374bbd"} Dec 01 08:59:08 crc kubenswrapper[4873]: I1201 08:59:08.203981 4873 generic.go:334] "Generic (PLEG): container finished" podID="570a0599-aafa-47fe-8b7a-26056d6a1723" containerID="5a5b3cbb7c950ec13da82b3af05ff052fa137e7cb753076bc7c3b959cf2e22bb" exitCode=0 Dec 01 08:59:08 crc kubenswrapper[4873]: I1201 08:59:08.204054 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-q6bwr" event={"ID":"570a0599-aafa-47fe-8b7a-26056d6a1723","Type":"ContainerDied","Data":"5a5b3cbb7c950ec13da82b3af05ff052fa137e7cb753076bc7c3b959cf2e22bb"} Dec 01 08:59:08 crc kubenswrapper[4873]: I1201 08:59:08.211851 4873 generic.go:334] "Generic (PLEG): container finished" podID="860ae34b-0993-4a11-ac3a-3774804a0860" containerID="08894b35673c62becdf6e89e1d7ceaba3675395a718195b458b0423be85453ac" exitCode=0 Dec 01 08:59:08 crc kubenswrapper[4873]: I1201 08:59:08.211955 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-h9jj9" event={"ID":"860ae34b-0993-4a11-ac3a-3774804a0860","Type":"ContainerDied","Data":"08894b35673c62becdf6e89e1d7ceaba3675395a718195b458b0423be85453ac"} Dec 01 08:59:08 crc kubenswrapper[4873]: I1201 08:59:08.214857 4873 generic.go:334] "Generic (PLEG): container finished" podID="a20240ed-c3f5-444d-854d-718d0f5cbace" containerID="243d39e2e5e4ccb73cb1c86fc2ff0271389dcbd153fc38dc44f58b96c2a27060" exitCode=0 Dec 01 08:59:08 crc kubenswrapper[4873]: I1201 08:59:08.214959 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-m9cfn" event={"ID":"a20240ed-c3f5-444d-854d-718d0f5cbace","Type":"ContainerDied","Data":"243d39e2e5e4ccb73cb1c86fc2ff0271389dcbd153fc38dc44f58b96c2a27060"} Dec 01 08:59:08 crc kubenswrapper[4873]: I1201 08:59:08.216887 4873 generic.go:334] "Generic (PLEG): container finished" podID="b5650152-030e-42d2-8db1-b80bfa18bc6f" containerID="67014367ee4a3300e046b29a2dbfb826636377bb9001a425a2675910de3226fa" exitCode=0 Dec 01 08:59:08 crc kubenswrapper[4873]: I1201 08:59:08.216957 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-108c-account-create-update-cmrw2" event={"ID":"b5650152-030e-42d2-8db1-b80bfa18bc6f","Type":"ContainerDied","Data":"67014367ee4a3300e046b29a2dbfb826636377bb9001a425a2675910de3226fa"} Dec 01 08:59:08 crc kubenswrapper[4873]: I1201 08:59:08.219233 4873 generic.go:334] "Generic (PLEG): container finished" podID="69e4256b-7143-4d91-91a0-b4ffdeb36511" containerID="6ac16135d2f590ecdcbb00ff124044df0d52cef64a72cf1df45041ba3f664814" exitCode=0 Dec 01 08:59:08 crc kubenswrapper[4873]: I1201 08:59:08.219308 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-3c32-account-create-update-vl8cf" event={"ID":"69e4256b-7143-4d91-91a0-b4ffdeb36511","Type":"ContainerDied","Data":"6ac16135d2f590ecdcbb00ff124044df0d52cef64a72cf1df45041ba3f664814"} Dec 01 08:59:08 crc kubenswrapper[4873]: I1201 08:59:08.322342 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54f9b7b8d9-t4rtz"] Dec 01 08:59:09 crc kubenswrapper[4873]: I1201 08:59:09.232810 4873 generic.go:334] "Generic (PLEG): container finished" podID="3c1fee8d-dc8f-4167-8d20-21ff76dd32de" containerID="0a167f1c59cd2ef54a4dee50fa3d2e584b21635e5a7cfc13cf2bea48e1132b42" exitCode=0 Dec 01 08:59:09 crc kubenswrapper[4873]: I1201 08:59:09.234232 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54f9b7b8d9-t4rtz" event={"ID":"3c1fee8d-dc8f-4167-8d20-21ff76dd32de","Type":"ContainerDied","Data":"0a167f1c59cd2ef54a4dee50fa3d2e584b21635e5a7cfc13cf2bea48e1132b42"} Dec 01 08:59:09 crc kubenswrapper[4873]: I1201 08:59:09.234318 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54f9b7b8d9-t4rtz" event={"ID":"3c1fee8d-dc8f-4167-8d20-21ff76dd32de","Type":"ContainerStarted","Data":"bb35b12d7a472bcbdc9544e9d227222eeea6c3188aafbbfcb038657f5128cbd5"} Dec 01 08:59:09 crc kubenswrapper[4873]: I1201 08:59:09.915107 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-eb5b-account-create-update-q6hf2" Dec 01 08:59:09 crc kubenswrapper[4873]: I1201 08:59:09.988134 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l52rq\" (UniqueName: \"kubernetes.io/projected/17c0c1a0-6746-4e83-aa1e-419bc703ac35-kube-api-access-l52rq\") pod \"17c0c1a0-6746-4e83-aa1e-419bc703ac35\" (UID: \"17c0c1a0-6746-4e83-aa1e-419bc703ac35\") " Dec 01 08:59:09 crc kubenswrapper[4873]: I1201 08:59:09.988328 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/17c0c1a0-6746-4e83-aa1e-419bc703ac35-operator-scripts\") pod \"17c0c1a0-6746-4e83-aa1e-419bc703ac35\" (UID: \"17c0c1a0-6746-4e83-aa1e-419bc703ac35\") " Dec 01 08:59:09 crc kubenswrapper[4873]: I1201 08:59:09.992874 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17c0c1a0-6746-4e83-aa1e-419bc703ac35-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "17c0c1a0-6746-4e83-aa1e-419bc703ac35" (UID: "17c0c1a0-6746-4e83-aa1e-419bc703ac35"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.018465 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17c0c1a0-6746-4e83-aa1e-419bc703ac35-kube-api-access-l52rq" (OuterVolumeSpecName: "kube-api-access-l52rq") pod "17c0c1a0-6746-4e83-aa1e-419bc703ac35" (UID: "17c0c1a0-6746-4e83-aa1e-419bc703ac35"). InnerVolumeSpecName "kube-api-access-l52rq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.075752 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-m9cfn" Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.085103 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-3c32-account-create-update-vl8cf" Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.092875 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a20240ed-c3f5-444d-854d-718d0f5cbace-operator-scripts\") pod \"a20240ed-c3f5-444d-854d-718d0f5cbace\" (UID: \"a20240ed-c3f5-444d-854d-718d0f5cbace\") " Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.092995 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/69e4256b-7143-4d91-91a0-b4ffdeb36511-operator-scripts\") pod \"69e4256b-7143-4d91-91a0-b4ffdeb36511\" (UID: \"69e4256b-7143-4d91-91a0-b4ffdeb36511\") " Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.093053 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ph6td\" (UniqueName: \"kubernetes.io/projected/a20240ed-c3f5-444d-854d-718d0f5cbace-kube-api-access-ph6td\") pod \"a20240ed-c3f5-444d-854d-718d0f5cbace\" (UID: \"a20240ed-c3f5-444d-854d-718d0f5cbace\") " Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.093156 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5f95\" (UniqueName: \"kubernetes.io/projected/69e4256b-7143-4d91-91a0-b4ffdeb36511-kube-api-access-f5f95\") pod \"69e4256b-7143-4d91-91a0-b4ffdeb36511\" (UID: \"69e4256b-7143-4d91-91a0-b4ffdeb36511\") " Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.093416 4873 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/17c0c1a0-6746-4e83-aa1e-419bc703ac35-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.093429 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l52rq\" (UniqueName: \"kubernetes.io/projected/17c0c1a0-6746-4e83-aa1e-419bc703ac35-kube-api-access-l52rq\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.093840 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-108c-account-create-update-cmrw2" Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.098315 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69e4256b-7143-4d91-91a0-b4ffdeb36511-kube-api-access-f5f95" (OuterVolumeSpecName: "kube-api-access-f5f95") pod "69e4256b-7143-4d91-91a0-b4ffdeb36511" (UID: "69e4256b-7143-4d91-91a0-b4ffdeb36511"). InnerVolumeSpecName "kube-api-access-f5f95". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.098691 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a20240ed-c3f5-444d-854d-718d0f5cbace-kube-api-access-ph6td" (OuterVolumeSpecName: "kube-api-access-ph6td") pod "a20240ed-c3f5-444d-854d-718d0f5cbace" (UID: "a20240ed-c3f5-444d-854d-718d0f5cbace"). InnerVolumeSpecName "kube-api-access-ph6td". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.117842 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-h9jj9" Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.194614 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/860ae34b-0993-4a11-ac3a-3774804a0860-operator-scripts\") pod \"860ae34b-0993-4a11-ac3a-3774804a0860\" (UID: \"860ae34b-0993-4a11-ac3a-3774804a0860\") " Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.194675 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5650152-030e-42d2-8db1-b80bfa18bc6f-operator-scripts\") pod \"b5650152-030e-42d2-8db1-b80bfa18bc6f\" (UID: \"b5650152-030e-42d2-8db1-b80bfa18bc6f\") " Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.194744 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-np66w\" (UniqueName: \"kubernetes.io/projected/860ae34b-0993-4a11-ac3a-3774804a0860-kube-api-access-np66w\") pod \"860ae34b-0993-4a11-ac3a-3774804a0860\" (UID: \"860ae34b-0993-4a11-ac3a-3774804a0860\") " Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.194815 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58mkh\" (UniqueName: \"kubernetes.io/projected/b5650152-030e-42d2-8db1-b80bfa18bc6f-kube-api-access-58mkh\") pod \"b5650152-030e-42d2-8db1-b80bfa18bc6f\" (UID: \"b5650152-030e-42d2-8db1-b80bfa18bc6f\") " Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.195109 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ph6td\" (UniqueName: \"kubernetes.io/projected/a20240ed-c3f5-444d-854d-718d0f5cbace-kube-api-access-ph6td\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.195126 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5f95\" (UniqueName: \"kubernetes.io/projected/69e4256b-7143-4d91-91a0-b4ffdeb36511-kube-api-access-f5f95\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.203712 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5650152-030e-42d2-8db1-b80bfa18bc6f-kube-api-access-58mkh" (OuterVolumeSpecName: "kube-api-access-58mkh") pod "b5650152-030e-42d2-8db1-b80bfa18bc6f" (UID: "b5650152-030e-42d2-8db1-b80bfa18bc6f"). InnerVolumeSpecName "kube-api-access-58mkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.203875 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/860ae34b-0993-4a11-ac3a-3774804a0860-kube-api-access-np66w" (OuterVolumeSpecName: "kube-api-access-np66w") pod "860ae34b-0993-4a11-ac3a-3774804a0860" (UID: "860ae34b-0993-4a11-ac3a-3774804a0860"). InnerVolumeSpecName "kube-api-access-np66w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.259063 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-108c-account-create-update-cmrw2" Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.259087 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-108c-account-create-update-cmrw2" event={"ID":"b5650152-030e-42d2-8db1-b80bfa18bc6f","Type":"ContainerDied","Data":"64d3f102ff960b2291f754d88e27ed61be361757a44096601615129b751f88f8"} Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.259311 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="64d3f102ff960b2291f754d88e27ed61be361757a44096601615129b751f88f8" Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.261367 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-3c32-account-create-update-vl8cf" event={"ID":"69e4256b-7143-4d91-91a0-b4ffdeb36511","Type":"ContainerDied","Data":"acaca650ecfe205574dc9441965a62e774e2bfd43e65ef530b86d70dcfb00a6c"} Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.261405 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="acaca650ecfe205574dc9441965a62e774e2bfd43e65ef530b86d70dcfb00a6c" Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.261482 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-3c32-account-create-update-vl8cf" Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.269720 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-eb5b-account-create-update-q6hf2" event={"ID":"17c0c1a0-6746-4e83-aa1e-419bc703ac35","Type":"ContainerDied","Data":"9566b575c98cb13799c4b3c5b7371a10491dd128e237a17fb45b37db5beea866"} Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.269761 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9566b575c98cb13799c4b3c5b7371a10491dd128e237a17fb45b37db5beea866" Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.269802 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-eb5b-account-create-update-q6hf2" Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.271572 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-q6bwr" Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.274971 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-h9jj9" event={"ID":"860ae34b-0993-4a11-ac3a-3774804a0860","Type":"ContainerDied","Data":"ecd836ffe8b4facb66082b9f3a97c9bec85993c050228026c2d30fe48f30eb9c"} Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.275000 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ecd836ffe8b4facb66082b9f3a97c9bec85993c050228026c2d30fe48f30eb9c" Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.275057 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-h9jj9" Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.277686 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-m9cfn" event={"ID":"a20240ed-c3f5-444d-854d-718d0f5cbace","Type":"ContainerDied","Data":"ea38facb65915324d92cd0d34edfbca41a5ae04dc7a0e50511be5da3dd4af76e"} Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.277739 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea38facb65915324d92cd0d34edfbca41a5ae04dc7a0e50511be5da3dd4af76e" Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.277744 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-m9cfn" Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.297615 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-np66w\" (UniqueName: \"kubernetes.io/projected/860ae34b-0993-4a11-ac3a-3774804a0860-kube-api-access-np66w\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.297969 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-58mkh\" (UniqueName: \"kubernetes.io/projected/b5650152-030e-42d2-8db1-b80bfa18bc6f-kube-api-access-58mkh\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.383571 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a20240ed-c3f5-444d-854d-718d0f5cbace-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a20240ed-c3f5-444d-854d-718d0f5cbace" (UID: "a20240ed-c3f5-444d-854d-718d0f5cbace"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.383618 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69e4256b-7143-4d91-91a0-b4ffdeb36511-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "69e4256b-7143-4d91-91a0-b4ffdeb36511" (UID: "69e4256b-7143-4d91-91a0-b4ffdeb36511"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.383712 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/860ae34b-0993-4a11-ac3a-3774804a0860-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "860ae34b-0993-4a11-ac3a-3774804a0860" (UID: "860ae34b-0993-4a11-ac3a-3774804a0860"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.383830 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5650152-030e-42d2-8db1-b80bfa18bc6f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b5650152-030e-42d2-8db1-b80bfa18bc6f" (UID: "b5650152-030e-42d2-8db1-b80bfa18bc6f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.399412 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvxpc\" (UniqueName: \"kubernetes.io/projected/570a0599-aafa-47fe-8b7a-26056d6a1723-kube-api-access-pvxpc\") pod \"570a0599-aafa-47fe-8b7a-26056d6a1723\" (UID: \"570a0599-aafa-47fe-8b7a-26056d6a1723\") " Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.399519 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/570a0599-aafa-47fe-8b7a-26056d6a1723-operator-scripts\") pod \"570a0599-aafa-47fe-8b7a-26056d6a1723\" (UID: \"570a0599-aafa-47fe-8b7a-26056d6a1723\") " Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.400063 4873 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a20240ed-c3f5-444d-854d-718d0f5cbace-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.400084 4873 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/860ae34b-0993-4a11-ac3a-3774804a0860-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.400098 4873 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5650152-030e-42d2-8db1-b80bfa18bc6f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.400109 4873 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/69e4256b-7143-4d91-91a0-b4ffdeb36511-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.400123 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/570a0599-aafa-47fe-8b7a-26056d6a1723-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "570a0599-aafa-47fe-8b7a-26056d6a1723" (UID: "570a0599-aafa-47fe-8b7a-26056d6a1723"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.407028 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/570a0599-aafa-47fe-8b7a-26056d6a1723-kube-api-access-pvxpc" (OuterVolumeSpecName: "kube-api-access-pvxpc") pod "570a0599-aafa-47fe-8b7a-26056d6a1723" (UID: "570a0599-aafa-47fe-8b7a-26056d6a1723"). InnerVolumeSpecName "kube-api-access-pvxpc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.501960 4873 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/570a0599-aafa-47fe-8b7a-26056d6a1723-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:10 crc kubenswrapper[4873]: I1201 08:59:10.502074 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvxpc\" (UniqueName: \"kubernetes.io/projected/570a0599-aafa-47fe-8b7a-26056d6a1723-kube-api-access-pvxpc\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:11 crc kubenswrapper[4873]: I1201 08:59:11.291151 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54f9b7b8d9-t4rtz" event={"ID":"3c1fee8d-dc8f-4167-8d20-21ff76dd32de","Type":"ContainerStarted","Data":"539f87537745220b07fb39faf639760032a7e74d7310a53693c1981fff1fc864"} Dec 01 08:59:11 crc kubenswrapper[4873]: I1201 08:59:11.291881 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-54f9b7b8d9-t4rtz" Dec 01 08:59:11 crc kubenswrapper[4873]: I1201 08:59:11.297234 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-q6bwr" event={"ID":"570a0599-aafa-47fe-8b7a-26056d6a1723","Type":"ContainerDied","Data":"37671b622cdb11993e87bc46503d759901179105f757d67329e96cc990a8c8d9"} Dec 01 08:59:11 crc kubenswrapper[4873]: I1201 08:59:11.297280 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="37671b622cdb11993e87bc46503d759901179105f757d67329e96cc990a8c8d9" Dec 01 08:59:11 crc kubenswrapper[4873]: I1201 08:59:11.297295 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-q6bwr" Dec 01 08:59:11 crc kubenswrapper[4873]: I1201 08:59:11.328005 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-54f9b7b8d9-t4rtz" podStartSLOduration=4.327980137 podStartE2EDuration="4.327980137s" podCreationTimestamp="2025-12-01 08:59:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:59:11.319715582 +0000 UTC m=+1127.221824131" watchObservedRunningTime="2025-12-01 08:59:11.327980137 +0000 UTC m=+1127.230088676" Dec 01 08:59:16 crc kubenswrapper[4873]: I1201 08:59:16.417829 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-wshpl" event={"ID":"1b86ab79-9562-43e9-a624-f0bc8e1e16c3","Type":"ContainerStarted","Data":"8afd0e65b2766a22e9a567e19eab5c347622e6d24d1fd8a897a5058866896007"} Dec 01 08:59:16 crc kubenswrapper[4873]: I1201 08:59:16.439131 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-wshpl" podStartSLOduration=2.343187081 podStartE2EDuration="11.43910319s" podCreationTimestamp="2025-12-01 08:59:05 +0000 UTC" firstStartedPulling="2025-12-01 08:59:06.503773728 +0000 UTC m=+1122.405882267" lastFinishedPulling="2025-12-01 08:59:15.599689837 +0000 UTC m=+1131.501798376" observedRunningTime="2025-12-01 08:59:16.433907237 +0000 UTC m=+1132.336015766" watchObservedRunningTime="2025-12-01 08:59:16.43910319 +0000 UTC m=+1132.341211719" Dec 01 08:59:17 crc kubenswrapper[4873]: I1201 08:59:17.826404 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-54f9b7b8d9-t4rtz" Dec 01 08:59:17 crc kubenswrapper[4873]: I1201 08:59:17.895349 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-j7n86"] Dec 01 08:59:17 crc kubenswrapper[4873]: I1201 08:59:17.902453 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-j7n86" podUID="11ec3abe-6052-445d-9f3f-016157fecbf9" containerName="dnsmasq-dns" containerID="cri-o://e8b70798f88aed17ff235aba1f24cb3657753f831be3e65ea9bb413ec4f1b38e" gracePeriod=10 Dec 01 08:59:18 crc kubenswrapper[4873]: I1201 08:59:18.477632 4873 generic.go:334] "Generic (PLEG): container finished" podID="11ec3abe-6052-445d-9f3f-016157fecbf9" containerID="e8b70798f88aed17ff235aba1f24cb3657753f831be3e65ea9bb413ec4f1b38e" exitCode=0 Dec 01 08:59:18 crc kubenswrapper[4873]: I1201 08:59:18.478960 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-j7n86" event={"ID":"11ec3abe-6052-445d-9f3f-016157fecbf9","Type":"ContainerDied","Data":"e8b70798f88aed17ff235aba1f24cb3657753f831be3e65ea9bb413ec4f1b38e"} Dec 01 08:59:18 crc kubenswrapper[4873]: I1201 08:59:18.638257 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-j7n86" Dec 01 08:59:18 crc kubenswrapper[4873]: I1201 08:59:18.824617 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11ec3abe-6052-445d-9f3f-016157fecbf9-dns-svc\") pod \"11ec3abe-6052-445d-9f3f-016157fecbf9\" (UID: \"11ec3abe-6052-445d-9f3f-016157fecbf9\") " Dec 01 08:59:18 crc kubenswrapper[4873]: I1201 08:59:18.824712 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/11ec3abe-6052-445d-9f3f-016157fecbf9-ovsdbserver-sb\") pod \"11ec3abe-6052-445d-9f3f-016157fecbf9\" (UID: \"11ec3abe-6052-445d-9f3f-016157fecbf9\") " Dec 01 08:59:18 crc kubenswrapper[4873]: I1201 08:59:18.824934 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11ec3abe-6052-445d-9f3f-016157fecbf9-ovsdbserver-nb\") pod \"11ec3abe-6052-445d-9f3f-016157fecbf9\" (UID: \"11ec3abe-6052-445d-9f3f-016157fecbf9\") " Dec 01 08:59:18 crc kubenswrapper[4873]: I1201 08:59:18.826224 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11ec3abe-6052-445d-9f3f-016157fecbf9-config\") pod \"11ec3abe-6052-445d-9f3f-016157fecbf9\" (UID: \"11ec3abe-6052-445d-9f3f-016157fecbf9\") " Dec 01 08:59:18 crc kubenswrapper[4873]: I1201 08:59:18.826333 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjpkk\" (UniqueName: \"kubernetes.io/projected/11ec3abe-6052-445d-9f3f-016157fecbf9-kube-api-access-pjpkk\") pod \"11ec3abe-6052-445d-9f3f-016157fecbf9\" (UID: \"11ec3abe-6052-445d-9f3f-016157fecbf9\") " Dec 01 08:59:18 crc kubenswrapper[4873]: I1201 08:59:18.833509 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11ec3abe-6052-445d-9f3f-016157fecbf9-kube-api-access-pjpkk" (OuterVolumeSpecName: "kube-api-access-pjpkk") pod "11ec3abe-6052-445d-9f3f-016157fecbf9" (UID: "11ec3abe-6052-445d-9f3f-016157fecbf9"). InnerVolumeSpecName "kube-api-access-pjpkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:59:18 crc kubenswrapper[4873]: I1201 08:59:18.879145 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11ec3abe-6052-445d-9f3f-016157fecbf9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "11ec3abe-6052-445d-9f3f-016157fecbf9" (UID: "11ec3abe-6052-445d-9f3f-016157fecbf9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:59:18 crc kubenswrapper[4873]: I1201 08:59:18.880724 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11ec3abe-6052-445d-9f3f-016157fecbf9-config" (OuterVolumeSpecName: "config") pod "11ec3abe-6052-445d-9f3f-016157fecbf9" (UID: "11ec3abe-6052-445d-9f3f-016157fecbf9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:59:18 crc kubenswrapper[4873]: I1201 08:59:18.883501 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11ec3abe-6052-445d-9f3f-016157fecbf9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "11ec3abe-6052-445d-9f3f-016157fecbf9" (UID: "11ec3abe-6052-445d-9f3f-016157fecbf9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:59:18 crc kubenswrapper[4873]: I1201 08:59:18.893044 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11ec3abe-6052-445d-9f3f-016157fecbf9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "11ec3abe-6052-445d-9f3f-016157fecbf9" (UID: "11ec3abe-6052-445d-9f3f-016157fecbf9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:59:18 crc kubenswrapper[4873]: I1201 08:59:18.930748 4873 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/11ec3abe-6052-445d-9f3f-016157fecbf9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:18 crc kubenswrapper[4873]: I1201 08:59:18.930789 4873 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11ec3abe-6052-445d-9f3f-016157fecbf9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:18 crc kubenswrapper[4873]: I1201 08:59:18.930801 4873 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11ec3abe-6052-445d-9f3f-016157fecbf9-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:18 crc kubenswrapper[4873]: I1201 08:59:18.930813 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjpkk\" (UniqueName: \"kubernetes.io/projected/11ec3abe-6052-445d-9f3f-016157fecbf9-kube-api-access-pjpkk\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:18 crc kubenswrapper[4873]: I1201 08:59:18.930830 4873 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11ec3abe-6052-445d-9f3f-016157fecbf9-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:19 crc kubenswrapper[4873]: I1201 08:59:19.489416 4873 generic.go:334] "Generic (PLEG): container finished" podID="1b86ab79-9562-43e9-a624-f0bc8e1e16c3" containerID="8afd0e65b2766a22e9a567e19eab5c347622e6d24d1fd8a897a5058866896007" exitCode=0 Dec 01 08:59:19 crc kubenswrapper[4873]: I1201 08:59:19.489498 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-wshpl" event={"ID":"1b86ab79-9562-43e9-a624-f0bc8e1e16c3","Type":"ContainerDied","Data":"8afd0e65b2766a22e9a567e19eab5c347622e6d24d1fd8a897a5058866896007"} Dec 01 08:59:19 crc kubenswrapper[4873]: I1201 08:59:19.491959 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-j7n86" event={"ID":"11ec3abe-6052-445d-9f3f-016157fecbf9","Type":"ContainerDied","Data":"28c37e91ea5eba47df338112136f5e1f44d396525851c5242ebca47b9ce4addd"} Dec 01 08:59:19 crc kubenswrapper[4873]: I1201 08:59:19.492034 4873 scope.go:117] "RemoveContainer" containerID="e8b70798f88aed17ff235aba1f24cb3657753f831be3e65ea9bb413ec4f1b38e" Dec 01 08:59:19 crc kubenswrapper[4873]: I1201 08:59:19.492038 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-j7n86" Dec 01 08:59:19 crc kubenswrapper[4873]: I1201 08:59:19.545204 4873 scope.go:117] "RemoveContainer" containerID="f3643b9dfddb59b751277cd5d49925dbeae682a21e8f4202abe8de17ee25358e" Dec 01 08:59:19 crc kubenswrapper[4873]: I1201 08:59:19.550844 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-j7n86"] Dec 01 08:59:19 crc kubenswrapper[4873]: I1201 08:59:19.561285 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-j7n86"] Dec 01 08:59:20 crc kubenswrapper[4873]: I1201 08:59:20.442365 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11ec3abe-6052-445d-9f3f-016157fecbf9" path="/var/lib/kubelet/pods/11ec3abe-6052-445d-9f3f-016157fecbf9/volumes" Dec 01 08:59:20 crc kubenswrapper[4873]: I1201 08:59:20.956085 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-wshpl" Dec 01 08:59:21 crc kubenswrapper[4873]: I1201 08:59:21.147927 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2kd4g\" (UniqueName: \"kubernetes.io/projected/1b86ab79-9562-43e9-a624-f0bc8e1e16c3-kube-api-access-2kd4g\") pod \"1b86ab79-9562-43e9-a624-f0bc8e1e16c3\" (UID: \"1b86ab79-9562-43e9-a624-f0bc8e1e16c3\") " Dec 01 08:59:21 crc kubenswrapper[4873]: I1201 08:59:21.148110 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b86ab79-9562-43e9-a624-f0bc8e1e16c3-config-data\") pod \"1b86ab79-9562-43e9-a624-f0bc8e1e16c3\" (UID: \"1b86ab79-9562-43e9-a624-f0bc8e1e16c3\") " Dec 01 08:59:21 crc kubenswrapper[4873]: I1201 08:59:21.148186 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b86ab79-9562-43e9-a624-f0bc8e1e16c3-combined-ca-bundle\") pod \"1b86ab79-9562-43e9-a624-f0bc8e1e16c3\" (UID: \"1b86ab79-9562-43e9-a624-f0bc8e1e16c3\") " Dec 01 08:59:21 crc kubenswrapper[4873]: I1201 08:59:21.162569 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b86ab79-9562-43e9-a624-f0bc8e1e16c3-kube-api-access-2kd4g" (OuterVolumeSpecName: "kube-api-access-2kd4g") pod "1b86ab79-9562-43e9-a624-f0bc8e1e16c3" (UID: "1b86ab79-9562-43e9-a624-f0bc8e1e16c3"). InnerVolumeSpecName "kube-api-access-2kd4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:59:21 crc kubenswrapper[4873]: I1201 08:59:21.175326 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b86ab79-9562-43e9-a624-f0bc8e1e16c3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1b86ab79-9562-43e9-a624-f0bc8e1e16c3" (UID: "1b86ab79-9562-43e9-a624-f0bc8e1e16c3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:59:21 crc kubenswrapper[4873]: I1201 08:59:21.194543 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b86ab79-9562-43e9-a624-f0bc8e1e16c3-config-data" (OuterVolumeSpecName: "config-data") pod "1b86ab79-9562-43e9-a624-f0bc8e1e16c3" (UID: "1b86ab79-9562-43e9-a624-f0bc8e1e16c3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:59:21 crc kubenswrapper[4873]: I1201 08:59:21.250866 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2kd4g\" (UniqueName: \"kubernetes.io/projected/1b86ab79-9562-43e9-a624-f0bc8e1e16c3-kube-api-access-2kd4g\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:21 crc kubenswrapper[4873]: I1201 08:59:21.250928 4873 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b86ab79-9562-43e9-a624-f0bc8e1e16c3-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:21 crc kubenswrapper[4873]: I1201 08:59:21.250939 4873 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b86ab79-9562-43e9-a624-f0bc8e1e16c3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:21 crc kubenswrapper[4873]: I1201 08:59:21.520944 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-wshpl" event={"ID":"1b86ab79-9562-43e9-a624-f0bc8e1e16c3","Type":"ContainerDied","Data":"3ba2708931673d7bd4dbc73d5c5d4359bceb072ebf22296660d7cad677f79dcb"} Dec 01 08:59:21 crc kubenswrapper[4873]: I1201 08:59:21.522834 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3ba2708931673d7bd4dbc73d5c5d4359bceb072ebf22296660d7cad677f79dcb" Dec 01 08:59:21 crc kubenswrapper[4873]: I1201 08:59:21.521089 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-wshpl" Dec 01 08:59:21 crc kubenswrapper[4873]: I1201 08:59:21.913297 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6546db6db7-wqvtd"] Dec 01 08:59:21 crc kubenswrapper[4873]: E1201 08:59:21.914411 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="570a0599-aafa-47fe-8b7a-26056d6a1723" containerName="mariadb-database-create" Dec 01 08:59:21 crc kubenswrapper[4873]: I1201 08:59:21.914433 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="570a0599-aafa-47fe-8b7a-26056d6a1723" containerName="mariadb-database-create" Dec 01 08:59:21 crc kubenswrapper[4873]: E1201 08:59:21.914454 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5650152-030e-42d2-8db1-b80bfa18bc6f" containerName="mariadb-account-create-update" Dec 01 08:59:21 crc kubenswrapper[4873]: I1201 08:59:21.914461 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5650152-030e-42d2-8db1-b80bfa18bc6f" containerName="mariadb-account-create-update" Dec 01 08:59:21 crc kubenswrapper[4873]: E1201 08:59:21.914480 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69e4256b-7143-4d91-91a0-b4ffdeb36511" containerName="mariadb-account-create-update" Dec 01 08:59:21 crc kubenswrapper[4873]: I1201 08:59:21.914489 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="69e4256b-7143-4d91-91a0-b4ffdeb36511" containerName="mariadb-account-create-update" Dec 01 08:59:21 crc kubenswrapper[4873]: E1201 08:59:21.914502 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17c0c1a0-6746-4e83-aa1e-419bc703ac35" containerName="mariadb-account-create-update" Dec 01 08:59:21 crc kubenswrapper[4873]: I1201 08:59:21.914509 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="17c0c1a0-6746-4e83-aa1e-419bc703ac35" containerName="mariadb-account-create-update" Dec 01 08:59:21 crc kubenswrapper[4873]: E1201 08:59:21.914528 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a20240ed-c3f5-444d-854d-718d0f5cbace" containerName="mariadb-database-create" Dec 01 08:59:21 crc kubenswrapper[4873]: I1201 08:59:21.914535 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="a20240ed-c3f5-444d-854d-718d0f5cbace" containerName="mariadb-database-create" Dec 01 08:59:21 crc kubenswrapper[4873]: E1201 08:59:21.914550 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11ec3abe-6052-445d-9f3f-016157fecbf9" containerName="dnsmasq-dns" Dec 01 08:59:21 crc kubenswrapper[4873]: I1201 08:59:21.914557 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="11ec3abe-6052-445d-9f3f-016157fecbf9" containerName="dnsmasq-dns" Dec 01 08:59:21 crc kubenswrapper[4873]: E1201 08:59:21.914566 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11ec3abe-6052-445d-9f3f-016157fecbf9" containerName="init" Dec 01 08:59:21 crc kubenswrapper[4873]: I1201 08:59:21.914573 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="11ec3abe-6052-445d-9f3f-016157fecbf9" containerName="init" Dec 01 08:59:21 crc kubenswrapper[4873]: E1201 08:59:21.914581 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b86ab79-9562-43e9-a624-f0bc8e1e16c3" containerName="keystone-db-sync" Dec 01 08:59:21 crc kubenswrapper[4873]: I1201 08:59:21.914588 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b86ab79-9562-43e9-a624-f0bc8e1e16c3" containerName="keystone-db-sync" Dec 01 08:59:21 crc kubenswrapper[4873]: E1201 08:59:21.914602 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="860ae34b-0993-4a11-ac3a-3774804a0860" containerName="mariadb-database-create" Dec 01 08:59:21 crc kubenswrapper[4873]: I1201 08:59:21.914608 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="860ae34b-0993-4a11-ac3a-3774804a0860" containerName="mariadb-database-create" Dec 01 08:59:21 crc kubenswrapper[4873]: I1201 08:59:21.914799 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="860ae34b-0993-4a11-ac3a-3774804a0860" containerName="mariadb-database-create" Dec 01 08:59:21 crc kubenswrapper[4873]: I1201 08:59:21.914847 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="17c0c1a0-6746-4e83-aa1e-419bc703ac35" containerName="mariadb-account-create-update" Dec 01 08:59:21 crc kubenswrapper[4873]: I1201 08:59:21.914857 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="a20240ed-c3f5-444d-854d-718d0f5cbace" containerName="mariadb-database-create" Dec 01 08:59:21 crc kubenswrapper[4873]: I1201 08:59:21.914866 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="69e4256b-7143-4d91-91a0-b4ffdeb36511" containerName="mariadb-account-create-update" Dec 01 08:59:21 crc kubenswrapper[4873]: I1201 08:59:21.914875 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b86ab79-9562-43e9-a624-f0bc8e1e16c3" containerName="keystone-db-sync" Dec 01 08:59:21 crc kubenswrapper[4873]: I1201 08:59:21.914887 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="570a0599-aafa-47fe-8b7a-26056d6a1723" containerName="mariadb-database-create" Dec 01 08:59:21 crc kubenswrapper[4873]: I1201 08:59:21.914898 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5650152-030e-42d2-8db1-b80bfa18bc6f" containerName="mariadb-account-create-update" Dec 01 08:59:21 crc kubenswrapper[4873]: I1201 08:59:21.914905 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="11ec3abe-6052-445d-9f3f-016157fecbf9" containerName="dnsmasq-dns" Dec 01 08:59:21 crc kubenswrapper[4873]: I1201 08:59:21.916291 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6546db6db7-wqvtd" Dec 01 08:59:21 crc kubenswrapper[4873]: I1201 08:59:21.941508 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6546db6db7-wqvtd"] Dec 01 08:59:21 crc kubenswrapper[4873]: I1201 08:59:21.977556 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/04cc97ba-b3a6-4906-87d1-b94c44858e86-ovsdbserver-nb\") pod \"dnsmasq-dns-6546db6db7-wqvtd\" (UID: \"04cc97ba-b3a6-4906-87d1-b94c44858e86\") " pod="openstack/dnsmasq-dns-6546db6db7-wqvtd" Dec 01 08:59:21 crc kubenswrapper[4873]: I1201 08:59:21.977634 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvxtt\" (UniqueName: \"kubernetes.io/projected/04cc97ba-b3a6-4906-87d1-b94c44858e86-kube-api-access-fvxtt\") pod \"dnsmasq-dns-6546db6db7-wqvtd\" (UID: \"04cc97ba-b3a6-4906-87d1-b94c44858e86\") " pod="openstack/dnsmasq-dns-6546db6db7-wqvtd" Dec 01 08:59:21 crc kubenswrapper[4873]: I1201 08:59:21.977693 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04cc97ba-b3a6-4906-87d1-b94c44858e86-config\") pod \"dnsmasq-dns-6546db6db7-wqvtd\" (UID: \"04cc97ba-b3a6-4906-87d1-b94c44858e86\") " pod="openstack/dnsmasq-dns-6546db6db7-wqvtd" Dec 01 08:59:21 crc kubenswrapper[4873]: I1201 08:59:21.977748 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/04cc97ba-b3a6-4906-87d1-b94c44858e86-dns-svc\") pod \"dnsmasq-dns-6546db6db7-wqvtd\" (UID: \"04cc97ba-b3a6-4906-87d1-b94c44858e86\") " pod="openstack/dnsmasq-dns-6546db6db7-wqvtd" Dec 01 08:59:21 crc kubenswrapper[4873]: I1201 08:59:21.977778 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/04cc97ba-b3a6-4906-87d1-b94c44858e86-ovsdbserver-sb\") pod \"dnsmasq-dns-6546db6db7-wqvtd\" (UID: \"04cc97ba-b3a6-4906-87d1-b94c44858e86\") " pod="openstack/dnsmasq-dns-6546db6db7-wqvtd" Dec 01 08:59:21 crc kubenswrapper[4873]: I1201 08:59:21.986075 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-7xln4"] Dec 01 08:59:21 crc kubenswrapper[4873]: I1201 08:59:21.987820 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-7xln4" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.000120 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.000363 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-9jwfx" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.000625 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.000782 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.008461 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.008667 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-7xln4"] Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.088901 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ac6fb56-fa10-4936-b8b3-1204d14945f8-scripts\") pod \"keystone-bootstrap-7xln4\" (UID: \"7ac6fb56-fa10-4936-b8b3-1204d14945f8\") " pod="openstack/keystone-bootstrap-7xln4" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.088956 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7ac6fb56-fa10-4936-b8b3-1204d14945f8-credential-keys\") pod \"keystone-bootstrap-7xln4\" (UID: \"7ac6fb56-fa10-4936-b8b3-1204d14945f8\") " pod="openstack/keystone-bootstrap-7xln4" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.089008 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ac6fb56-fa10-4936-b8b3-1204d14945f8-combined-ca-bundle\") pod \"keystone-bootstrap-7xln4\" (UID: \"7ac6fb56-fa10-4936-b8b3-1204d14945f8\") " pod="openstack/keystone-bootstrap-7xln4" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.089069 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/04cc97ba-b3a6-4906-87d1-b94c44858e86-dns-svc\") pod \"dnsmasq-dns-6546db6db7-wqvtd\" (UID: \"04cc97ba-b3a6-4906-87d1-b94c44858e86\") " pod="openstack/dnsmasq-dns-6546db6db7-wqvtd" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.089096 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpk8v\" (UniqueName: \"kubernetes.io/projected/7ac6fb56-fa10-4936-b8b3-1204d14945f8-kube-api-access-dpk8v\") pod \"keystone-bootstrap-7xln4\" (UID: \"7ac6fb56-fa10-4936-b8b3-1204d14945f8\") " pod="openstack/keystone-bootstrap-7xln4" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.089116 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/04cc97ba-b3a6-4906-87d1-b94c44858e86-ovsdbserver-sb\") pod \"dnsmasq-dns-6546db6db7-wqvtd\" (UID: \"04cc97ba-b3a6-4906-87d1-b94c44858e86\") " pod="openstack/dnsmasq-dns-6546db6db7-wqvtd" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.089137 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ac6fb56-fa10-4936-b8b3-1204d14945f8-config-data\") pod \"keystone-bootstrap-7xln4\" (UID: \"7ac6fb56-fa10-4936-b8b3-1204d14945f8\") " pod="openstack/keystone-bootstrap-7xln4" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.089154 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/04cc97ba-b3a6-4906-87d1-b94c44858e86-ovsdbserver-nb\") pod \"dnsmasq-dns-6546db6db7-wqvtd\" (UID: \"04cc97ba-b3a6-4906-87d1-b94c44858e86\") " pod="openstack/dnsmasq-dns-6546db6db7-wqvtd" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.089193 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvxtt\" (UniqueName: \"kubernetes.io/projected/04cc97ba-b3a6-4906-87d1-b94c44858e86-kube-api-access-fvxtt\") pod \"dnsmasq-dns-6546db6db7-wqvtd\" (UID: \"04cc97ba-b3a6-4906-87d1-b94c44858e86\") " pod="openstack/dnsmasq-dns-6546db6db7-wqvtd" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.089240 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7ac6fb56-fa10-4936-b8b3-1204d14945f8-fernet-keys\") pod \"keystone-bootstrap-7xln4\" (UID: \"7ac6fb56-fa10-4936-b8b3-1204d14945f8\") " pod="openstack/keystone-bootstrap-7xln4" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.089266 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04cc97ba-b3a6-4906-87d1-b94c44858e86-config\") pod \"dnsmasq-dns-6546db6db7-wqvtd\" (UID: \"04cc97ba-b3a6-4906-87d1-b94c44858e86\") " pod="openstack/dnsmasq-dns-6546db6db7-wqvtd" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.090515 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04cc97ba-b3a6-4906-87d1-b94c44858e86-config\") pod \"dnsmasq-dns-6546db6db7-wqvtd\" (UID: \"04cc97ba-b3a6-4906-87d1-b94c44858e86\") " pod="openstack/dnsmasq-dns-6546db6db7-wqvtd" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.092183 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/04cc97ba-b3a6-4906-87d1-b94c44858e86-dns-svc\") pod \"dnsmasq-dns-6546db6db7-wqvtd\" (UID: \"04cc97ba-b3a6-4906-87d1-b94c44858e86\") " pod="openstack/dnsmasq-dns-6546db6db7-wqvtd" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.092378 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/04cc97ba-b3a6-4906-87d1-b94c44858e86-ovsdbserver-nb\") pod \"dnsmasq-dns-6546db6db7-wqvtd\" (UID: \"04cc97ba-b3a6-4906-87d1-b94c44858e86\") " pod="openstack/dnsmasq-dns-6546db6db7-wqvtd" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.092824 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/04cc97ba-b3a6-4906-87d1-b94c44858e86-ovsdbserver-sb\") pod \"dnsmasq-dns-6546db6db7-wqvtd\" (UID: \"04cc97ba-b3a6-4906-87d1-b94c44858e86\") " pod="openstack/dnsmasq-dns-6546db6db7-wqvtd" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.120566 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-bmhcv"] Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.122596 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-bmhcv" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.124131 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvxtt\" (UniqueName: \"kubernetes.io/projected/04cc97ba-b3a6-4906-87d1-b94c44858e86-kube-api-access-fvxtt\") pod \"dnsmasq-dns-6546db6db7-wqvtd\" (UID: \"04cc97ba-b3a6-4906-87d1-b94c44858e86\") " pod="openstack/dnsmasq-dns-6546db6db7-wqvtd" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.131677 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-snh6v" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.137896 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.141850 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.143619 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-bmhcv"] Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.228730 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ac6fb56-fa10-4936-b8b3-1204d14945f8-config-data\") pod \"keystone-bootstrap-7xln4\" (UID: \"7ac6fb56-fa10-4936-b8b3-1204d14945f8\") " pod="openstack/keystone-bootstrap-7xln4" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.229045 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7ac6fb56-fa10-4936-b8b3-1204d14945f8-fernet-keys\") pod \"keystone-bootstrap-7xln4\" (UID: \"7ac6fb56-fa10-4936-b8b3-1204d14945f8\") " pod="openstack/keystone-bootstrap-7xln4" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.229144 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ac6fb56-fa10-4936-b8b3-1204d14945f8-scripts\") pod \"keystone-bootstrap-7xln4\" (UID: \"7ac6fb56-fa10-4936-b8b3-1204d14945f8\") " pod="openstack/keystone-bootstrap-7xln4" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.229169 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7ac6fb56-fa10-4936-b8b3-1204d14945f8-credential-keys\") pod \"keystone-bootstrap-7xln4\" (UID: \"7ac6fb56-fa10-4936-b8b3-1204d14945f8\") " pod="openstack/keystone-bootstrap-7xln4" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.229276 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ac6fb56-fa10-4936-b8b3-1204d14945f8-combined-ca-bundle\") pod \"keystone-bootstrap-7xln4\" (UID: \"7ac6fb56-fa10-4936-b8b3-1204d14945f8\") " pod="openstack/keystone-bootstrap-7xln4" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.229390 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpk8v\" (UniqueName: \"kubernetes.io/projected/7ac6fb56-fa10-4936-b8b3-1204d14945f8-kube-api-access-dpk8v\") pod \"keystone-bootstrap-7xln4\" (UID: \"7ac6fb56-fa10-4936-b8b3-1204d14945f8\") " pod="openstack/keystone-bootstrap-7xln4" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.239726 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ac6fb56-fa10-4936-b8b3-1204d14945f8-combined-ca-bundle\") pod \"keystone-bootstrap-7xln4\" (UID: \"7ac6fb56-fa10-4936-b8b3-1204d14945f8\") " pod="openstack/keystone-bootstrap-7xln4" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.243773 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ac6fb56-fa10-4936-b8b3-1204d14945f8-scripts\") pod \"keystone-bootstrap-7xln4\" (UID: \"7ac6fb56-fa10-4936-b8b3-1204d14945f8\") " pod="openstack/keystone-bootstrap-7xln4" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.250414 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6546db6db7-wqvtd" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.258750 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7ac6fb56-fa10-4936-b8b3-1204d14945f8-credential-keys\") pod \"keystone-bootstrap-7xln4\" (UID: \"7ac6fb56-fa10-4936-b8b3-1204d14945f8\") " pod="openstack/keystone-bootstrap-7xln4" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.266933 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ac6fb56-fa10-4936-b8b3-1204d14945f8-config-data\") pod \"keystone-bootstrap-7xln4\" (UID: \"7ac6fb56-fa10-4936-b8b3-1204d14945f8\") " pod="openstack/keystone-bootstrap-7xln4" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.281902 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7ac6fb56-fa10-4936-b8b3-1204d14945f8-fernet-keys\") pod \"keystone-bootstrap-7xln4\" (UID: \"7ac6fb56-fa10-4936-b8b3-1204d14945f8\") " pod="openstack/keystone-bootstrap-7xln4" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.293833 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpk8v\" (UniqueName: \"kubernetes.io/projected/7ac6fb56-fa10-4936-b8b3-1204d14945f8-kube-api-access-dpk8v\") pod \"keystone-bootstrap-7xln4\" (UID: \"7ac6fb56-fa10-4936-b8b3-1204d14945f8\") " pod="openstack/keystone-bootstrap-7xln4" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.293973 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-5vlz5"] Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.299705 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-5vlz5" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.303834 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-92gc9" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.304284 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.304541 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.327811 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-5vlz5"] Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.331382 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtl44\" (UniqueName: \"kubernetes.io/projected/ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce-kube-api-access-rtl44\") pod \"cinder-db-sync-bmhcv\" (UID: \"ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce\") " pod="openstack/cinder-db-sync-bmhcv" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.331517 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce-scripts\") pod \"cinder-db-sync-bmhcv\" (UID: \"ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce\") " pod="openstack/cinder-db-sync-bmhcv" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.331575 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce-config-data\") pod \"cinder-db-sync-bmhcv\" (UID: \"ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce\") " pod="openstack/cinder-db-sync-bmhcv" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.331610 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce-etc-machine-id\") pod \"cinder-db-sync-bmhcv\" (UID: \"ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce\") " pod="openstack/cinder-db-sync-bmhcv" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.331640 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce-db-sync-config-data\") pod \"cinder-db-sync-bmhcv\" (UID: \"ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce\") " pod="openstack/cinder-db-sync-bmhcv" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.331665 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce-combined-ca-bundle\") pod \"cinder-db-sync-bmhcv\" (UID: \"ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce\") " pod="openstack/cinder-db-sync-bmhcv" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.342738 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-7xln4" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.367565 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-9lvtd"] Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.369173 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-9lvtd" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.376657 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-qmv66" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.378982 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.401911 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-9lvtd"] Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.434205 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce-scripts\") pod \"cinder-db-sync-bmhcv\" (UID: \"ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce\") " pod="openstack/cinder-db-sync-bmhcv" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.437498 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkmlc\" (UniqueName: \"kubernetes.io/projected/ec6a6ffa-2ab6-41c4-b594-17066d11654a-kube-api-access-qkmlc\") pod \"neutron-db-sync-5vlz5\" (UID: \"ec6a6ffa-2ab6-41c4-b594-17066d11654a\") " pod="openstack/neutron-db-sync-5vlz5" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.437606 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce-config-data\") pod \"cinder-db-sync-bmhcv\" (UID: \"ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce\") " pod="openstack/cinder-db-sync-bmhcv" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.437633 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce-etc-machine-id\") pod \"cinder-db-sync-bmhcv\" (UID: \"ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce\") " pod="openstack/cinder-db-sync-bmhcv" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.437716 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce-db-sync-config-data\") pod \"cinder-db-sync-bmhcv\" (UID: \"ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce\") " pod="openstack/cinder-db-sync-bmhcv" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.437753 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce-combined-ca-bundle\") pod \"cinder-db-sync-bmhcv\" (UID: \"ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce\") " pod="openstack/cinder-db-sync-bmhcv" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.437866 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ec6a6ffa-2ab6-41c4-b594-17066d11654a-config\") pod \"neutron-db-sync-5vlz5\" (UID: \"ec6a6ffa-2ab6-41c4-b594-17066d11654a\") " pod="openstack/neutron-db-sync-5vlz5" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.437894 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtl44\" (UniqueName: \"kubernetes.io/projected/ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce-kube-api-access-rtl44\") pod \"cinder-db-sync-bmhcv\" (UID: \"ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce\") " pod="openstack/cinder-db-sync-bmhcv" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.437912 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec6a6ffa-2ab6-41c4-b594-17066d11654a-combined-ca-bundle\") pod \"neutron-db-sync-5vlz5\" (UID: \"ec6a6ffa-2ab6-41c4-b594-17066d11654a\") " pod="openstack/neutron-db-sync-5vlz5" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.450288 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce-etc-machine-id\") pod \"cinder-db-sync-bmhcv\" (UID: \"ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce\") " pod="openstack/cinder-db-sync-bmhcv" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.451946 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce-scripts\") pod \"cinder-db-sync-bmhcv\" (UID: \"ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce\") " pod="openstack/cinder-db-sync-bmhcv" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.453555 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce-config-data\") pod \"cinder-db-sync-bmhcv\" (UID: \"ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce\") " pod="openstack/cinder-db-sync-bmhcv" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.473680 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce-combined-ca-bundle\") pod \"cinder-db-sync-bmhcv\" (UID: \"ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce\") " pod="openstack/cinder-db-sync-bmhcv" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.473784 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce-db-sync-config-data\") pod \"cinder-db-sync-bmhcv\" (UID: \"ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce\") " pod="openstack/cinder-db-sync-bmhcv" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.517777 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtl44\" (UniqueName: \"kubernetes.io/projected/ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce-kube-api-access-rtl44\") pod \"cinder-db-sync-bmhcv\" (UID: \"ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce\") " pod="openstack/cinder-db-sync-bmhcv" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.529912 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-bmhcv" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.530512 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-nmmcg"] Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.532052 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-nmmcg" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.539746 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-fgj2w" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.539989 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.540202 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.547080 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c8c5adca-9738-46b1-9376-01cfb425d839-db-sync-config-data\") pod \"barbican-db-sync-9lvtd\" (UID: \"c8c5adca-9738-46b1-9376-01cfb425d839\") " pod="openstack/barbican-db-sync-9lvtd" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.547157 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkmlc\" (UniqueName: \"kubernetes.io/projected/ec6a6ffa-2ab6-41c4-b594-17066d11654a-kube-api-access-qkmlc\") pod \"neutron-db-sync-5vlz5\" (UID: \"ec6a6ffa-2ab6-41c4-b594-17066d11654a\") " pod="openstack/neutron-db-sync-5vlz5" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.547308 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ec6a6ffa-2ab6-41c4-b594-17066d11654a-config\") pod \"neutron-db-sync-5vlz5\" (UID: \"ec6a6ffa-2ab6-41c4-b594-17066d11654a\") " pod="openstack/neutron-db-sync-5vlz5" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.547333 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec6a6ffa-2ab6-41c4-b594-17066d11654a-combined-ca-bundle\") pod \"neutron-db-sync-5vlz5\" (UID: \"ec6a6ffa-2ab6-41c4-b594-17066d11654a\") " pod="openstack/neutron-db-sync-5vlz5" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.547393 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h64g6\" (UniqueName: \"kubernetes.io/projected/c8c5adca-9738-46b1-9376-01cfb425d839-kube-api-access-h64g6\") pod \"barbican-db-sync-9lvtd\" (UID: \"c8c5adca-9738-46b1-9376-01cfb425d839\") " pod="openstack/barbican-db-sync-9lvtd" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.547460 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8c5adca-9738-46b1-9376-01cfb425d839-combined-ca-bundle\") pod \"barbican-db-sync-9lvtd\" (UID: \"c8c5adca-9738-46b1-9376-01cfb425d839\") " pod="openstack/barbican-db-sync-9lvtd" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.554913 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/ec6a6ffa-2ab6-41c4-b594-17066d11654a-config\") pod \"neutron-db-sync-5vlz5\" (UID: \"ec6a6ffa-2ab6-41c4-b594-17066d11654a\") " pod="openstack/neutron-db-sync-5vlz5" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.560042 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec6a6ffa-2ab6-41c4-b594-17066d11654a-combined-ca-bundle\") pod \"neutron-db-sync-5vlz5\" (UID: \"ec6a6ffa-2ab6-41c4-b594-17066d11654a\") " pod="openstack/neutron-db-sync-5vlz5" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.571113 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-nmmcg"] Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.589363 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkmlc\" (UniqueName: \"kubernetes.io/projected/ec6a6ffa-2ab6-41c4-b594-17066d11654a-kube-api-access-qkmlc\") pod \"neutron-db-sync-5vlz5\" (UID: \"ec6a6ffa-2ab6-41c4-b594-17066d11654a\") " pod="openstack/neutron-db-sync-5vlz5" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.607748 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6546db6db7-wqvtd"] Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.644926 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.648677 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.649679 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0351e0d-24dc-407a-ba64-a38b313cf17c-config-data\") pod \"placement-db-sync-nmmcg\" (UID: \"b0351e0d-24dc-407a-ba64-a38b313cf17c\") " pod="openstack/placement-db-sync-nmmcg" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.649825 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0351e0d-24dc-407a-ba64-a38b313cf17c-logs\") pod \"placement-db-sync-nmmcg\" (UID: \"b0351e0d-24dc-407a-ba64-a38b313cf17c\") " pod="openstack/placement-db-sync-nmmcg" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.649860 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h64g6\" (UniqueName: \"kubernetes.io/projected/c8c5adca-9738-46b1-9376-01cfb425d839-kube-api-access-h64g6\") pod \"barbican-db-sync-9lvtd\" (UID: \"c8c5adca-9738-46b1-9376-01cfb425d839\") " pod="openstack/barbican-db-sync-9lvtd" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.649886 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b0351e0d-24dc-407a-ba64-a38b313cf17c-scripts\") pod \"placement-db-sync-nmmcg\" (UID: \"b0351e0d-24dc-407a-ba64-a38b313cf17c\") " pod="openstack/placement-db-sync-nmmcg" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.649908 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjqdf\" (UniqueName: \"kubernetes.io/projected/b0351e0d-24dc-407a-ba64-a38b313cf17c-kube-api-access-xjqdf\") pod \"placement-db-sync-nmmcg\" (UID: \"b0351e0d-24dc-407a-ba64-a38b313cf17c\") " pod="openstack/placement-db-sync-nmmcg" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.649945 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0351e0d-24dc-407a-ba64-a38b313cf17c-combined-ca-bundle\") pod \"placement-db-sync-nmmcg\" (UID: \"b0351e0d-24dc-407a-ba64-a38b313cf17c\") " pod="openstack/placement-db-sync-nmmcg" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.649968 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8c5adca-9738-46b1-9376-01cfb425d839-combined-ca-bundle\") pod \"barbican-db-sync-9lvtd\" (UID: \"c8c5adca-9738-46b1-9376-01cfb425d839\") " pod="openstack/barbican-db-sync-9lvtd" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.649992 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c8c5adca-9738-46b1-9376-01cfb425d839-db-sync-config-data\") pod \"barbican-db-sync-9lvtd\" (UID: \"c8c5adca-9738-46b1-9376-01cfb425d839\") " pod="openstack/barbican-db-sync-9lvtd" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.652500 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.652792 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.653209 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.660649 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8c5adca-9738-46b1-9376-01cfb425d839-combined-ca-bundle\") pod \"barbican-db-sync-9lvtd\" (UID: \"c8c5adca-9738-46b1-9376-01cfb425d839\") " pod="openstack/barbican-db-sync-9lvtd" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.666990 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7987f74bbc-6rfmd"] Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.679170 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c8c5adca-9738-46b1-9376-01cfb425d839-db-sync-config-data\") pod \"barbican-db-sync-9lvtd\" (UID: \"c8c5adca-9738-46b1-9376-01cfb425d839\") " pod="openstack/barbican-db-sync-9lvtd" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.684465 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7987f74bbc-6rfmd" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.685542 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h64g6\" (UniqueName: \"kubernetes.io/projected/c8c5adca-9738-46b1-9376-01cfb425d839-kube-api-access-h64g6\") pod \"barbican-db-sync-9lvtd\" (UID: \"c8c5adca-9738-46b1-9376-01cfb425d839\") " pod="openstack/barbican-db-sync-9lvtd" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.724619 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7987f74bbc-6rfmd"] Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.737378 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-5vlz5" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.751799 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e53689a0-5d92-4e62-bb63-3a56af762a59-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e53689a0-5d92-4e62-bb63-3a56af762a59\") " pod="openstack/ceilometer-0" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.751864 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b0351e0d-24dc-407a-ba64-a38b313cf17c-scripts\") pod \"placement-db-sync-nmmcg\" (UID: \"b0351e0d-24dc-407a-ba64-a38b313cf17c\") " pod="openstack/placement-db-sync-nmmcg" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.751892 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjqdf\" (UniqueName: \"kubernetes.io/projected/b0351e0d-24dc-407a-ba64-a38b313cf17c-kube-api-access-xjqdf\") pod \"placement-db-sync-nmmcg\" (UID: \"b0351e0d-24dc-407a-ba64-a38b313cf17c\") " pod="openstack/placement-db-sync-nmmcg" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.751912 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e53689a0-5d92-4e62-bb63-3a56af762a59-scripts\") pod \"ceilometer-0\" (UID: \"e53689a0-5d92-4e62-bb63-3a56af762a59\") " pod="openstack/ceilometer-0" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.751948 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0351e0d-24dc-407a-ba64-a38b313cf17c-combined-ca-bundle\") pod \"placement-db-sync-nmmcg\" (UID: \"b0351e0d-24dc-407a-ba64-a38b313cf17c\") " pod="openstack/placement-db-sync-nmmcg" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.751982 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e53689a0-5d92-4e62-bb63-3a56af762a59-run-httpd\") pod \"ceilometer-0\" (UID: \"e53689a0-5d92-4e62-bb63-3a56af762a59\") " pod="openstack/ceilometer-0" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.752008 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e53689a0-5d92-4e62-bb63-3a56af762a59-config-data\") pod \"ceilometer-0\" (UID: \"e53689a0-5d92-4e62-bb63-3a56af762a59\") " pod="openstack/ceilometer-0" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.752056 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0351e0d-24dc-407a-ba64-a38b313cf17c-config-data\") pod \"placement-db-sync-nmmcg\" (UID: \"b0351e0d-24dc-407a-ba64-a38b313cf17c\") " pod="openstack/placement-db-sync-nmmcg" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.752083 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e53689a0-5d92-4e62-bb63-3a56af762a59-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e53689a0-5d92-4e62-bb63-3a56af762a59\") " pod="openstack/ceilometer-0" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.752109 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e53689a0-5d92-4e62-bb63-3a56af762a59-log-httpd\") pod \"ceilometer-0\" (UID: \"e53689a0-5d92-4e62-bb63-3a56af762a59\") " pod="openstack/ceilometer-0" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.752141 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spv5x\" (UniqueName: \"kubernetes.io/projected/e53689a0-5d92-4e62-bb63-3a56af762a59-kube-api-access-spv5x\") pod \"ceilometer-0\" (UID: \"e53689a0-5d92-4e62-bb63-3a56af762a59\") " pod="openstack/ceilometer-0" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.752203 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0351e0d-24dc-407a-ba64-a38b313cf17c-logs\") pod \"placement-db-sync-nmmcg\" (UID: \"b0351e0d-24dc-407a-ba64-a38b313cf17c\") " pod="openstack/placement-db-sync-nmmcg" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.754770 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0351e0d-24dc-407a-ba64-a38b313cf17c-logs\") pod \"placement-db-sync-nmmcg\" (UID: \"b0351e0d-24dc-407a-ba64-a38b313cf17c\") " pod="openstack/placement-db-sync-nmmcg" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.761972 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0351e0d-24dc-407a-ba64-a38b313cf17c-config-data\") pod \"placement-db-sync-nmmcg\" (UID: \"b0351e0d-24dc-407a-ba64-a38b313cf17c\") " pod="openstack/placement-db-sync-nmmcg" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.761976 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b0351e0d-24dc-407a-ba64-a38b313cf17c-scripts\") pod \"placement-db-sync-nmmcg\" (UID: \"b0351e0d-24dc-407a-ba64-a38b313cf17c\") " pod="openstack/placement-db-sync-nmmcg" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.772163 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0351e0d-24dc-407a-ba64-a38b313cf17c-combined-ca-bundle\") pod \"placement-db-sync-nmmcg\" (UID: \"b0351e0d-24dc-407a-ba64-a38b313cf17c\") " pod="openstack/placement-db-sync-nmmcg" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.783816 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjqdf\" (UniqueName: \"kubernetes.io/projected/b0351e0d-24dc-407a-ba64-a38b313cf17c-kube-api-access-xjqdf\") pod \"placement-db-sync-nmmcg\" (UID: \"b0351e0d-24dc-407a-ba64-a38b313cf17c\") " pod="openstack/placement-db-sync-nmmcg" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.806339 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-9lvtd" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.853963 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spv5x\" (UniqueName: \"kubernetes.io/projected/e53689a0-5d92-4e62-bb63-3a56af762a59-kube-api-access-spv5x\") pod \"ceilometer-0\" (UID: \"e53689a0-5d92-4e62-bb63-3a56af762a59\") " pod="openstack/ceilometer-0" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.854071 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njv2r\" (UniqueName: \"kubernetes.io/projected/4f027d3a-e35f-418b-bc8b-508a637f6698-kube-api-access-njv2r\") pod \"dnsmasq-dns-7987f74bbc-6rfmd\" (UID: \"4f027d3a-e35f-418b-bc8b-508a637f6698\") " pod="openstack/dnsmasq-dns-7987f74bbc-6rfmd" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.854102 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f027d3a-e35f-418b-bc8b-508a637f6698-ovsdbserver-nb\") pod \"dnsmasq-dns-7987f74bbc-6rfmd\" (UID: \"4f027d3a-e35f-418b-bc8b-508a637f6698\") " pod="openstack/dnsmasq-dns-7987f74bbc-6rfmd" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.854161 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e53689a0-5d92-4e62-bb63-3a56af762a59-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e53689a0-5d92-4e62-bb63-3a56af762a59\") " pod="openstack/ceilometer-0" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.854184 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4f027d3a-e35f-418b-bc8b-508a637f6698-ovsdbserver-sb\") pod \"dnsmasq-dns-7987f74bbc-6rfmd\" (UID: \"4f027d3a-e35f-418b-bc8b-508a637f6698\") " pod="openstack/dnsmasq-dns-7987f74bbc-6rfmd" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.854206 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f027d3a-e35f-418b-bc8b-508a637f6698-dns-svc\") pod \"dnsmasq-dns-7987f74bbc-6rfmd\" (UID: \"4f027d3a-e35f-418b-bc8b-508a637f6698\") " pod="openstack/dnsmasq-dns-7987f74bbc-6rfmd" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.854225 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e53689a0-5d92-4e62-bb63-3a56af762a59-scripts\") pod \"ceilometer-0\" (UID: \"e53689a0-5d92-4e62-bb63-3a56af762a59\") " pod="openstack/ceilometer-0" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.854276 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e53689a0-5d92-4e62-bb63-3a56af762a59-run-httpd\") pod \"ceilometer-0\" (UID: \"e53689a0-5d92-4e62-bb63-3a56af762a59\") " pod="openstack/ceilometer-0" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.854311 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e53689a0-5d92-4e62-bb63-3a56af762a59-config-data\") pod \"ceilometer-0\" (UID: \"e53689a0-5d92-4e62-bb63-3a56af762a59\") " pod="openstack/ceilometer-0" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.854325 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f027d3a-e35f-418b-bc8b-508a637f6698-config\") pod \"dnsmasq-dns-7987f74bbc-6rfmd\" (UID: \"4f027d3a-e35f-418b-bc8b-508a637f6698\") " pod="openstack/dnsmasq-dns-7987f74bbc-6rfmd" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.854364 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e53689a0-5d92-4e62-bb63-3a56af762a59-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e53689a0-5d92-4e62-bb63-3a56af762a59\") " pod="openstack/ceilometer-0" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.854386 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e53689a0-5d92-4e62-bb63-3a56af762a59-log-httpd\") pod \"ceilometer-0\" (UID: \"e53689a0-5d92-4e62-bb63-3a56af762a59\") " pod="openstack/ceilometer-0" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.854584 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-nmmcg" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.855001 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e53689a0-5d92-4e62-bb63-3a56af762a59-log-httpd\") pod \"ceilometer-0\" (UID: \"e53689a0-5d92-4e62-bb63-3a56af762a59\") " pod="openstack/ceilometer-0" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.855830 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e53689a0-5d92-4e62-bb63-3a56af762a59-run-httpd\") pod \"ceilometer-0\" (UID: \"e53689a0-5d92-4e62-bb63-3a56af762a59\") " pod="openstack/ceilometer-0" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.858681 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e53689a0-5d92-4e62-bb63-3a56af762a59-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e53689a0-5d92-4e62-bb63-3a56af762a59\") " pod="openstack/ceilometer-0" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.859431 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e53689a0-5d92-4e62-bb63-3a56af762a59-scripts\") pod \"ceilometer-0\" (UID: \"e53689a0-5d92-4e62-bb63-3a56af762a59\") " pod="openstack/ceilometer-0" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.862389 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e53689a0-5d92-4e62-bb63-3a56af762a59-config-data\") pod \"ceilometer-0\" (UID: \"e53689a0-5d92-4e62-bb63-3a56af762a59\") " pod="openstack/ceilometer-0" Dec 01 08:59:22 crc kubenswrapper[4873]: I1201 08:59:22.862783 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e53689a0-5d92-4e62-bb63-3a56af762a59-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e53689a0-5d92-4e62-bb63-3a56af762a59\") " pod="openstack/ceilometer-0" Dec 01 08:59:23 crc kubenswrapper[4873]: I1201 08:59:22.960232 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njv2r\" (UniqueName: \"kubernetes.io/projected/4f027d3a-e35f-418b-bc8b-508a637f6698-kube-api-access-njv2r\") pod \"dnsmasq-dns-7987f74bbc-6rfmd\" (UID: \"4f027d3a-e35f-418b-bc8b-508a637f6698\") " pod="openstack/dnsmasq-dns-7987f74bbc-6rfmd" Dec 01 08:59:23 crc kubenswrapper[4873]: I1201 08:59:22.960318 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f027d3a-e35f-418b-bc8b-508a637f6698-ovsdbserver-nb\") pod \"dnsmasq-dns-7987f74bbc-6rfmd\" (UID: \"4f027d3a-e35f-418b-bc8b-508a637f6698\") " pod="openstack/dnsmasq-dns-7987f74bbc-6rfmd" Dec 01 08:59:23 crc kubenswrapper[4873]: I1201 08:59:22.960447 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4f027d3a-e35f-418b-bc8b-508a637f6698-ovsdbserver-sb\") pod \"dnsmasq-dns-7987f74bbc-6rfmd\" (UID: \"4f027d3a-e35f-418b-bc8b-508a637f6698\") " pod="openstack/dnsmasq-dns-7987f74bbc-6rfmd" Dec 01 08:59:23 crc kubenswrapper[4873]: I1201 08:59:22.960501 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f027d3a-e35f-418b-bc8b-508a637f6698-dns-svc\") pod \"dnsmasq-dns-7987f74bbc-6rfmd\" (UID: \"4f027d3a-e35f-418b-bc8b-508a637f6698\") " pod="openstack/dnsmasq-dns-7987f74bbc-6rfmd" Dec 01 08:59:23 crc kubenswrapper[4873]: I1201 08:59:22.960632 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f027d3a-e35f-418b-bc8b-508a637f6698-config\") pod \"dnsmasq-dns-7987f74bbc-6rfmd\" (UID: \"4f027d3a-e35f-418b-bc8b-508a637f6698\") " pod="openstack/dnsmasq-dns-7987f74bbc-6rfmd" Dec 01 08:59:23 crc kubenswrapper[4873]: I1201 08:59:23.405481 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f027d3a-e35f-418b-bc8b-508a637f6698-ovsdbserver-nb\") pod \"dnsmasq-dns-7987f74bbc-6rfmd\" (UID: \"4f027d3a-e35f-418b-bc8b-508a637f6698\") " pod="openstack/dnsmasq-dns-7987f74bbc-6rfmd" Dec 01 08:59:23 crc kubenswrapper[4873]: I1201 08:59:23.422866 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spv5x\" (UniqueName: \"kubernetes.io/projected/e53689a0-5d92-4e62-bb63-3a56af762a59-kube-api-access-spv5x\") pod \"ceilometer-0\" (UID: \"e53689a0-5d92-4e62-bb63-3a56af762a59\") " pod="openstack/ceilometer-0" Dec 01 08:59:23 crc kubenswrapper[4873]: I1201 08:59:23.442833 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6546db6db7-wqvtd"] Dec 01 08:59:23 crc kubenswrapper[4873]: I1201 08:59:23.460271 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4f027d3a-e35f-418b-bc8b-508a637f6698-ovsdbserver-sb\") pod \"dnsmasq-dns-7987f74bbc-6rfmd\" (UID: \"4f027d3a-e35f-418b-bc8b-508a637f6698\") " pod="openstack/dnsmasq-dns-7987f74bbc-6rfmd" Dec 01 08:59:23 crc kubenswrapper[4873]: I1201 08:59:23.463111 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f027d3a-e35f-418b-bc8b-508a637f6698-config\") pod \"dnsmasq-dns-7987f74bbc-6rfmd\" (UID: \"4f027d3a-e35f-418b-bc8b-508a637f6698\") " pod="openstack/dnsmasq-dns-7987f74bbc-6rfmd" Dec 01 08:59:23 crc kubenswrapper[4873]: I1201 08:59:23.458642 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f027d3a-e35f-418b-bc8b-508a637f6698-dns-svc\") pod \"dnsmasq-dns-7987f74bbc-6rfmd\" (UID: \"4f027d3a-e35f-418b-bc8b-508a637f6698\") " pod="openstack/dnsmasq-dns-7987f74bbc-6rfmd" Dec 01 08:59:23 crc kubenswrapper[4873]: I1201 08:59:23.498603 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njv2r\" (UniqueName: \"kubernetes.io/projected/4f027d3a-e35f-418b-bc8b-508a637f6698-kube-api-access-njv2r\") pod \"dnsmasq-dns-7987f74bbc-6rfmd\" (UID: \"4f027d3a-e35f-418b-bc8b-508a637f6698\") " pod="openstack/dnsmasq-dns-7987f74bbc-6rfmd" Dec 01 08:59:23 crc kubenswrapper[4873]: I1201 08:59:23.594916 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6546db6db7-wqvtd" event={"ID":"04cc97ba-b3a6-4906-87d1-b94c44858e86","Type":"ContainerStarted","Data":"2bd95c577dbfcffe25db2194e87c208c4d58f74bbf50889188084fa23282b756"} Dec 01 08:59:23 crc kubenswrapper[4873]: I1201 08:59:23.639820 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 08:59:23 crc kubenswrapper[4873]: I1201 08:59:23.654522 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7987f74bbc-6rfmd" Dec 01 08:59:23 crc kubenswrapper[4873]: I1201 08:59:23.890488 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-7xln4"] Dec 01 08:59:23 crc kubenswrapper[4873]: I1201 08:59:23.959722 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-bmhcv"] Dec 01 08:59:24 crc kubenswrapper[4873]: I1201 08:59:24.210052 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-5vlz5"] Dec 01 08:59:24 crc kubenswrapper[4873]: W1201 08:59:24.222651 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podec6a6ffa_2ab6_41c4_b594_17066d11654a.slice/crio-3b3b15f9d1116acf24cf5f574e1c3e9cadd04a88ee84b3b58e3be1104f470fb4 WatchSource:0}: Error finding container 3b3b15f9d1116acf24cf5f574e1c3e9cadd04a88ee84b3b58e3be1104f470fb4: Status 404 returned error can't find the container with id 3b3b15f9d1116acf24cf5f574e1c3e9cadd04a88ee84b3b58e3be1104f470fb4 Dec 01 08:59:24 crc kubenswrapper[4873]: I1201 08:59:24.244004 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-nmmcg"] Dec 01 08:59:24 crc kubenswrapper[4873]: I1201 08:59:24.262957 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-9lvtd"] Dec 01 08:59:24 crc kubenswrapper[4873]: W1201 08:59:24.493192 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f027d3a_e35f_418b_bc8b_508a637f6698.slice/crio-c752c8f201dda9c6baeeb62e7ac13a45eaf39e4b2157221faff6eb0f7b243d74 WatchSource:0}: Error finding container c752c8f201dda9c6baeeb62e7ac13a45eaf39e4b2157221faff6eb0f7b243d74: Status 404 returned error can't find the container with id c752c8f201dda9c6baeeb62e7ac13a45eaf39e4b2157221faff6eb0f7b243d74 Dec 01 08:59:24 crc kubenswrapper[4873]: I1201 08:59:24.558224 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7987f74bbc-6rfmd"] Dec 01 08:59:24 crc kubenswrapper[4873]: I1201 08:59:24.580199 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:59:24 crc kubenswrapper[4873]: I1201 08:59:24.616553 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-5vlz5" event={"ID":"ec6a6ffa-2ab6-41c4-b594-17066d11654a","Type":"ContainerStarted","Data":"3b3b15f9d1116acf24cf5f574e1c3e9cadd04a88ee84b3b58e3be1104f470fb4"} Dec 01 08:59:24 crc kubenswrapper[4873]: I1201 08:59:24.626560 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-nmmcg" event={"ID":"b0351e0d-24dc-407a-ba64-a38b313cf17c","Type":"ContainerStarted","Data":"16815ff095924e4ac372b522bcd3a5cbc6d13aec2fd2cc6d7fa125c7dc4e9b74"} Dec 01 08:59:24 crc kubenswrapper[4873]: I1201 08:59:24.628260 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-bmhcv" event={"ID":"ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce","Type":"ContainerStarted","Data":"17d905efe968146eb4a57a4bd628436f696ff2dd64309260dce5d653dde313b6"} Dec 01 08:59:24 crc kubenswrapper[4873]: I1201 08:59:24.640182 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e53689a0-5d92-4e62-bb63-3a56af762a59","Type":"ContainerStarted","Data":"98ad9d5515c70fcee8f018fb5f0cfdb6c0e0b5d6fcc8e7d6ed9338d994e68774"} Dec 01 08:59:24 crc kubenswrapper[4873]: I1201 08:59:24.647231 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7987f74bbc-6rfmd" event={"ID":"4f027d3a-e35f-418b-bc8b-508a637f6698","Type":"ContainerStarted","Data":"c752c8f201dda9c6baeeb62e7ac13a45eaf39e4b2157221faff6eb0f7b243d74"} Dec 01 08:59:24 crc kubenswrapper[4873]: I1201 08:59:24.649379 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-9lvtd" event={"ID":"c8c5adca-9738-46b1-9376-01cfb425d839","Type":"ContainerStarted","Data":"cec67fe9cd51f0192a8e60ce1cea570ee2e34eaa68a29440df00512e6246210f"} Dec 01 08:59:24 crc kubenswrapper[4873]: I1201 08:59:24.650911 4873 generic.go:334] "Generic (PLEG): container finished" podID="04cc97ba-b3a6-4906-87d1-b94c44858e86" containerID="7a35b9ecff6a76e39e21961a9d6688646404e6f0b545782f5ae0f3e5eacec31a" exitCode=0 Dec 01 08:59:24 crc kubenswrapper[4873]: I1201 08:59:24.650970 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6546db6db7-wqvtd" event={"ID":"04cc97ba-b3a6-4906-87d1-b94c44858e86","Type":"ContainerDied","Data":"7a35b9ecff6a76e39e21961a9d6688646404e6f0b545782f5ae0f3e5eacec31a"} Dec 01 08:59:24 crc kubenswrapper[4873]: I1201 08:59:24.656875 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-7xln4" event={"ID":"7ac6fb56-fa10-4936-b8b3-1204d14945f8","Type":"ContainerStarted","Data":"8290cdaf32cb1fefd00caf645207c20e538df3ee40573819f2a2ee25f0f677e5"} Dec 01 08:59:24 crc kubenswrapper[4873]: I1201 08:59:24.656940 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-7xln4" event={"ID":"7ac6fb56-fa10-4936-b8b3-1204d14945f8","Type":"ContainerStarted","Data":"3bdc010a2435f5982adb0308f29d079304c3381a7d22202d56b2e4d27447fe74"} Dec 01 08:59:24 crc kubenswrapper[4873]: I1201 08:59:24.755685 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-7xln4" podStartSLOduration=3.75563217 podStartE2EDuration="3.75563217s" podCreationTimestamp="2025-12-01 08:59:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:59:24.704957565 +0000 UTC m=+1140.607066094" watchObservedRunningTime="2025-12-01 08:59:24.75563217 +0000 UTC m=+1140.657740709" Dec 01 08:59:24 crc kubenswrapper[4873]: I1201 08:59:24.994554 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6546db6db7-wqvtd" Dec 01 08:59:25 crc kubenswrapper[4873]: I1201 08:59:25.139222 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/04cc97ba-b3a6-4906-87d1-b94c44858e86-dns-svc\") pod \"04cc97ba-b3a6-4906-87d1-b94c44858e86\" (UID: \"04cc97ba-b3a6-4906-87d1-b94c44858e86\") " Dec 01 08:59:25 crc kubenswrapper[4873]: I1201 08:59:25.139577 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fvxtt\" (UniqueName: \"kubernetes.io/projected/04cc97ba-b3a6-4906-87d1-b94c44858e86-kube-api-access-fvxtt\") pod \"04cc97ba-b3a6-4906-87d1-b94c44858e86\" (UID: \"04cc97ba-b3a6-4906-87d1-b94c44858e86\") " Dec 01 08:59:25 crc kubenswrapper[4873]: I1201 08:59:25.139679 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04cc97ba-b3a6-4906-87d1-b94c44858e86-config\") pod \"04cc97ba-b3a6-4906-87d1-b94c44858e86\" (UID: \"04cc97ba-b3a6-4906-87d1-b94c44858e86\") " Dec 01 08:59:25 crc kubenswrapper[4873]: I1201 08:59:25.139793 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/04cc97ba-b3a6-4906-87d1-b94c44858e86-ovsdbserver-nb\") pod \"04cc97ba-b3a6-4906-87d1-b94c44858e86\" (UID: \"04cc97ba-b3a6-4906-87d1-b94c44858e86\") " Dec 01 08:59:25 crc kubenswrapper[4873]: I1201 08:59:25.140072 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/04cc97ba-b3a6-4906-87d1-b94c44858e86-ovsdbserver-sb\") pod \"04cc97ba-b3a6-4906-87d1-b94c44858e86\" (UID: \"04cc97ba-b3a6-4906-87d1-b94c44858e86\") " Dec 01 08:59:25 crc kubenswrapper[4873]: I1201 08:59:25.145090 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04cc97ba-b3a6-4906-87d1-b94c44858e86-kube-api-access-fvxtt" (OuterVolumeSpecName: "kube-api-access-fvxtt") pod "04cc97ba-b3a6-4906-87d1-b94c44858e86" (UID: "04cc97ba-b3a6-4906-87d1-b94c44858e86"). InnerVolumeSpecName "kube-api-access-fvxtt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:59:25 crc kubenswrapper[4873]: I1201 08:59:25.171950 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04cc97ba-b3a6-4906-87d1-b94c44858e86-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "04cc97ba-b3a6-4906-87d1-b94c44858e86" (UID: "04cc97ba-b3a6-4906-87d1-b94c44858e86"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:59:25 crc kubenswrapper[4873]: I1201 08:59:25.181547 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04cc97ba-b3a6-4906-87d1-b94c44858e86-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "04cc97ba-b3a6-4906-87d1-b94c44858e86" (UID: "04cc97ba-b3a6-4906-87d1-b94c44858e86"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:59:25 crc kubenswrapper[4873]: I1201 08:59:25.197402 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04cc97ba-b3a6-4906-87d1-b94c44858e86-config" (OuterVolumeSpecName: "config") pod "04cc97ba-b3a6-4906-87d1-b94c44858e86" (UID: "04cc97ba-b3a6-4906-87d1-b94c44858e86"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:59:25 crc kubenswrapper[4873]: I1201 08:59:25.197570 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04cc97ba-b3a6-4906-87d1-b94c44858e86-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "04cc97ba-b3a6-4906-87d1-b94c44858e86" (UID: "04cc97ba-b3a6-4906-87d1-b94c44858e86"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:59:25 crc kubenswrapper[4873]: I1201 08:59:25.243560 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fvxtt\" (UniqueName: \"kubernetes.io/projected/04cc97ba-b3a6-4906-87d1-b94c44858e86-kube-api-access-fvxtt\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:25 crc kubenswrapper[4873]: I1201 08:59:25.243610 4873 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04cc97ba-b3a6-4906-87d1-b94c44858e86-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:25 crc kubenswrapper[4873]: I1201 08:59:25.243622 4873 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/04cc97ba-b3a6-4906-87d1-b94c44858e86-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:25 crc kubenswrapper[4873]: I1201 08:59:25.243631 4873 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/04cc97ba-b3a6-4906-87d1-b94c44858e86-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:25 crc kubenswrapper[4873]: I1201 08:59:25.243641 4873 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/04cc97ba-b3a6-4906-87d1-b94c44858e86-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:25 crc kubenswrapper[4873]: I1201 08:59:25.287451 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:59:25 crc kubenswrapper[4873]: I1201 08:59:25.681284 4873 generic.go:334] "Generic (PLEG): container finished" podID="4f027d3a-e35f-418b-bc8b-508a637f6698" containerID="90f9025bf705ac3103b5bb02995016ce74415078a98c44fd7a6326f903c2b040" exitCode=0 Dec 01 08:59:25 crc kubenswrapper[4873]: I1201 08:59:25.681357 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7987f74bbc-6rfmd" event={"ID":"4f027d3a-e35f-418b-bc8b-508a637f6698","Type":"ContainerDied","Data":"90f9025bf705ac3103b5bb02995016ce74415078a98c44fd7a6326f903c2b040"} Dec 01 08:59:25 crc kubenswrapper[4873]: I1201 08:59:25.690245 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6546db6db7-wqvtd" event={"ID":"04cc97ba-b3a6-4906-87d1-b94c44858e86","Type":"ContainerDied","Data":"2bd95c577dbfcffe25db2194e87c208c4d58f74bbf50889188084fa23282b756"} Dec 01 08:59:25 crc kubenswrapper[4873]: I1201 08:59:25.690307 4873 scope.go:117] "RemoveContainer" containerID="7a35b9ecff6a76e39e21961a9d6688646404e6f0b545782f5ae0f3e5eacec31a" Dec 01 08:59:25 crc kubenswrapper[4873]: I1201 08:59:25.690490 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6546db6db7-wqvtd" Dec 01 08:59:25 crc kubenswrapper[4873]: I1201 08:59:25.694690 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-5vlz5" event={"ID":"ec6a6ffa-2ab6-41c4-b594-17066d11654a","Type":"ContainerStarted","Data":"df48d3b6b18f16bbf985f991b069677905d37acec7b9825b5ae4eb31fc298878"} Dec 01 08:59:25 crc kubenswrapper[4873]: I1201 08:59:25.806760 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-5vlz5" podStartSLOduration=3.806738724 podStartE2EDuration="3.806738724s" podCreationTimestamp="2025-12-01 08:59:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:59:25.768200965 +0000 UTC m=+1141.670309514" watchObservedRunningTime="2025-12-01 08:59:25.806738724 +0000 UTC m=+1141.708847263" Dec 01 08:59:26 crc kubenswrapper[4873]: I1201 08:59:26.039866 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6546db6db7-wqvtd"] Dec 01 08:59:26 crc kubenswrapper[4873]: I1201 08:59:26.045564 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6546db6db7-wqvtd"] Dec 01 08:59:26 crc kubenswrapper[4873]: I1201 08:59:26.453679 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04cc97ba-b3a6-4906-87d1-b94c44858e86" path="/var/lib/kubelet/pods/04cc97ba-b3a6-4906-87d1-b94c44858e86/volumes" Dec 01 08:59:26 crc kubenswrapper[4873]: I1201 08:59:26.717153 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7987f74bbc-6rfmd" event={"ID":"4f027d3a-e35f-418b-bc8b-508a637f6698","Type":"ContainerStarted","Data":"63a7ee8c78eab560e2b238cb614c249216883b57b716a3838d82ba9f212e2127"} Dec 01 08:59:26 crc kubenswrapper[4873]: I1201 08:59:26.717594 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7987f74bbc-6rfmd" Dec 01 08:59:26 crc kubenswrapper[4873]: I1201 08:59:26.745851 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7987f74bbc-6rfmd" podStartSLOduration=4.745827606 podStartE2EDuration="4.745827606s" podCreationTimestamp="2025-12-01 08:59:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:59:26.736336242 +0000 UTC m=+1142.638444801" watchObservedRunningTime="2025-12-01 08:59:26.745827606 +0000 UTC m=+1142.647936145" Dec 01 08:59:29 crc kubenswrapper[4873]: I1201 08:59:29.757148 4873 generic.go:334] "Generic (PLEG): container finished" podID="7ac6fb56-fa10-4936-b8b3-1204d14945f8" containerID="8290cdaf32cb1fefd00caf645207c20e538df3ee40573819f2a2ee25f0f677e5" exitCode=0 Dec 01 08:59:29 crc kubenswrapper[4873]: I1201 08:59:29.757327 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-7xln4" event={"ID":"7ac6fb56-fa10-4936-b8b3-1204d14945f8","Type":"ContainerDied","Data":"8290cdaf32cb1fefd00caf645207c20e538df3ee40573819f2a2ee25f0f677e5"} Dec 01 08:59:31 crc kubenswrapper[4873]: I1201 08:59:31.059942 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:59:31 crc kubenswrapper[4873]: I1201 08:59:31.060532 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:59:31 crc kubenswrapper[4873]: I1201 08:59:31.060601 4873 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" Dec 01 08:59:31 crc kubenswrapper[4873]: I1201 08:59:31.061273 4873 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0976ae6cae359bbc0353d195c023558e0d48d039655f0dbf281e76be957bb9b6"} pod="openshift-machine-config-operator/machine-config-daemon-scwpp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 08:59:31 crc kubenswrapper[4873]: I1201 08:59:31.061332 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" containerID="cri-o://0976ae6cae359bbc0353d195c023558e0d48d039655f0dbf281e76be957bb9b6" gracePeriod=600 Dec 01 08:59:31 crc kubenswrapper[4873]: I1201 08:59:31.779209 4873 generic.go:334] "Generic (PLEG): container finished" podID="fef7b114-0e07-402d-a37b-315c36011f4b" containerID="0976ae6cae359bbc0353d195c023558e0d48d039655f0dbf281e76be957bb9b6" exitCode=0 Dec 01 08:59:31 crc kubenswrapper[4873]: I1201 08:59:31.779275 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" event={"ID":"fef7b114-0e07-402d-a37b-315c36011f4b","Type":"ContainerDied","Data":"0976ae6cae359bbc0353d195c023558e0d48d039655f0dbf281e76be957bb9b6"} Dec 01 08:59:31 crc kubenswrapper[4873]: I1201 08:59:31.779560 4873 scope.go:117] "RemoveContainer" containerID="2af36981211968ab2b3e6fc266a51d247e83e9c0c856f6566cda62f7eeb85b9a" Dec 01 08:59:33 crc kubenswrapper[4873]: I1201 08:59:33.789940 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7987f74bbc-6rfmd" Dec 01 08:59:33 crc kubenswrapper[4873]: I1201 08:59:33.866314 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54f9b7b8d9-t4rtz"] Dec 01 08:59:33 crc kubenswrapper[4873]: I1201 08:59:33.866708 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-54f9b7b8d9-t4rtz" podUID="3c1fee8d-dc8f-4167-8d20-21ff76dd32de" containerName="dnsmasq-dns" containerID="cri-o://539f87537745220b07fb39faf639760032a7e74d7310a53693c1981fff1fc864" gracePeriod=10 Dec 01 08:59:34 crc kubenswrapper[4873]: I1201 08:59:34.652129 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-7xln4" Dec 01 08:59:34 crc kubenswrapper[4873]: I1201 08:59:34.735406 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7ac6fb56-fa10-4936-b8b3-1204d14945f8-fernet-keys\") pod \"7ac6fb56-fa10-4936-b8b3-1204d14945f8\" (UID: \"7ac6fb56-fa10-4936-b8b3-1204d14945f8\") " Dec 01 08:59:34 crc kubenswrapper[4873]: I1201 08:59:34.735545 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ac6fb56-fa10-4936-b8b3-1204d14945f8-combined-ca-bundle\") pod \"7ac6fb56-fa10-4936-b8b3-1204d14945f8\" (UID: \"7ac6fb56-fa10-4936-b8b3-1204d14945f8\") " Dec 01 08:59:34 crc kubenswrapper[4873]: I1201 08:59:34.735755 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dpk8v\" (UniqueName: \"kubernetes.io/projected/7ac6fb56-fa10-4936-b8b3-1204d14945f8-kube-api-access-dpk8v\") pod \"7ac6fb56-fa10-4936-b8b3-1204d14945f8\" (UID: \"7ac6fb56-fa10-4936-b8b3-1204d14945f8\") " Dec 01 08:59:34 crc kubenswrapper[4873]: I1201 08:59:34.735795 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7ac6fb56-fa10-4936-b8b3-1204d14945f8-credential-keys\") pod \"7ac6fb56-fa10-4936-b8b3-1204d14945f8\" (UID: \"7ac6fb56-fa10-4936-b8b3-1204d14945f8\") " Dec 01 08:59:34 crc kubenswrapper[4873]: I1201 08:59:34.735823 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ac6fb56-fa10-4936-b8b3-1204d14945f8-config-data\") pod \"7ac6fb56-fa10-4936-b8b3-1204d14945f8\" (UID: \"7ac6fb56-fa10-4936-b8b3-1204d14945f8\") " Dec 01 08:59:34 crc kubenswrapper[4873]: I1201 08:59:34.735897 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ac6fb56-fa10-4936-b8b3-1204d14945f8-scripts\") pod \"7ac6fb56-fa10-4936-b8b3-1204d14945f8\" (UID: \"7ac6fb56-fa10-4936-b8b3-1204d14945f8\") " Dec 01 08:59:34 crc kubenswrapper[4873]: I1201 08:59:34.743172 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ac6fb56-fa10-4936-b8b3-1204d14945f8-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "7ac6fb56-fa10-4936-b8b3-1204d14945f8" (UID: "7ac6fb56-fa10-4936-b8b3-1204d14945f8"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:59:34 crc kubenswrapper[4873]: I1201 08:59:34.743341 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ac6fb56-fa10-4936-b8b3-1204d14945f8-scripts" (OuterVolumeSpecName: "scripts") pod "7ac6fb56-fa10-4936-b8b3-1204d14945f8" (UID: "7ac6fb56-fa10-4936-b8b3-1204d14945f8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:59:34 crc kubenswrapper[4873]: I1201 08:59:34.746712 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ac6fb56-fa10-4936-b8b3-1204d14945f8-kube-api-access-dpk8v" (OuterVolumeSpecName: "kube-api-access-dpk8v") pod "7ac6fb56-fa10-4936-b8b3-1204d14945f8" (UID: "7ac6fb56-fa10-4936-b8b3-1204d14945f8"). InnerVolumeSpecName "kube-api-access-dpk8v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:59:34 crc kubenswrapper[4873]: I1201 08:59:34.770671 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ac6fb56-fa10-4936-b8b3-1204d14945f8-config-data" (OuterVolumeSpecName: "config-data") pod "7ac6fb56-fa10-4936-b8b3-1204d14945f8" (UID: "7ac6fb56-fa10-4936-b8b3-1204d14945f8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:59:34 crc kubenswrapper[4873]: I1201 08:59:34.776540 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ac6fb56-fa10-4936-b8b3-1204d14945f8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7ac6fb56-fa10-4936-b8b3-1204d14945f8" (UID: "7ac6fb56-fa10-4936-b8b3-1204d14945f8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:59:34 crc kubenswrapper[4873]: I1201 08:59:34.782106 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ac6fb56-fa10-4936-b8b3-1204d14945f8-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "7ac6fb56-fa10-4936-b8b3-1204d14945f8" (UID: "7ac6fb56-fa10-4936-b8b3-1204d14945f8"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:59:34 crc kubenswrapper[4873]: I1201 08:59:34.829884 4873 generic.go:334] "Generic (PLEG): container finished" podID="3c1fee8d-dc8f-4167-8d20-21ff76dd32de" containerID="539f87537745220b07fb39faf639760032a7e74d7310a53693c1981fff1fc864" exitCode=0 Dec 01 08:59:34 crc kubenswrapper[4873]: I1201 08:59:34.830022 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54f9b7b8d9-t4rtz" event={"ID":"3c1fee8d-dc8f-4167-8d20-21ff76dd32de","Type":"ContainerDied","Data":"539f87537745220b07fb39faf639760032a7e74d7310a53693c1981fff1fc864"} Dec 01 08:59:34 crc kubenswrapper[4873]: I1201 08:59:34.839940 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-7xln4" event={"ID":"7ac6fb56-fa10-4936-b8b3-1204d14945f8","Type":"ContainerDied","Data":"3bdc010a2435f5982adb0308f29d079304c3381a7d22202d56b2e4d27447fe74"} Dec 01 08:59:34 crc kubenswrapper[4873]: I1201 08:59:34.839988 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3bdc010a2435f5982adb0308f29d079304c3381a7d22202d56b2e4d27447fe74" Dec 01 08:59:34 crc kubenswrapper[4873]: I1201 08:59:34.840123 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-7xln4" Dec 01 08:59:34 crc kubenswrapper[4873]: I1201 08:59:34.840917 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dpk8v\" (UniqueName: \"kubernetes.io/projected/7ac6fb56-fa10-4936-b8b3-1204d14945f8-kube-api-access-dpk8v\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:34 crc kubenswrapper[4873]: I1201 08:59:34.840961 4873 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7ac6fb56-fa10-4936-b8b3-1204d14945f8-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:34 crc kubenswrapper[4873]: I1201 08:59:34.840975 4873 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ac6fb56-fa10-4936-b8b3-1204d14945f8-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:34 crc kubenswrapper[4873]: I1201 08:59:34.840995 4873 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ac6fb56-fa10-4936-b8b3-1204d14945f8-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:34 crc kubenswrapper[4873]: I1201 08:59:34.841007 4873 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7ac6fb56-fa10-4936-b8b3-1204d14945f8-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:34 crc kubenswrapper[4873]: I1201 08:59:34.841063 4873 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ac6fb56-fa10-4936-b8b3-1204d14945f8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:35 crc kubenswrapper[4873]: I1201 08:59:35.144657 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54f9b7b8d9-t4rtz" Dec 01 08:59:35 crc kubenswrapper[4873]: I1201 08:59:35.246437 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c1fee8d-dc8f-4167-8d20-21ff76dd32de-config\") pod \"3c1fee8d-dc8f-4167-8d20-21ff76dd32de\" (UID: \"3c1fee8d-dc8f-4167-8d20-21ff76dd32de\") " Dec 01 08:59:35 crc kubenswrapper[4873]: I1201 08:59:35.248422 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bkz7k\" (UniqueName: \"kubernetes.io/projected/3c1fee8d-dc8f-4167-8d20-21ff76dd32de-kube-api-access-bkz7k\") pod \"3c1fee8d-dc8f-4167-8d20-21ff76dd32de\" (UID: \"3c1fee8d-dc8f-4167-8d20-21ff76dd32de\") " Dec 01 08:59:35 crc kubenswrapper[4873]: I1201 08:59:35.248656 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3c1fee8d-dc8f-4167-8d20-21ff76dd32de-ovsdbserver-sb\") pod \"3c1fee8d-dc8f-4167-8d20-21ff76dd32de\" (UID: \"3c1fee8d-dc8f-4167-8d20-21ff76dd32de\") " Dec 01 08:59:35 crc kubenswrapper[4873]: I1201 08:59:35.248753 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3c1fee8d-dc8f-4167-8d20-21ff76dd32de-ovsdbserver-nb\") pod \"3c1fee8d-dc8f-4167-8d20-21ff76dd32de\" (UID: \"3c1fee8d-dc8f-4167-8d20-21ff76dd32de\") " Dec 01 08:59:35 crc kubenswrapper[4873]: I1201 08:59:35.248800 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3c1fee8d-dc8f-4167-8d20-21ff76dd32de-dns-svc\") pod \"3c1fee8d-dc8f-4167-8d20-21ff76dd32de\" (UID: \"3c1fee8d-dc8f-4167-8d20-21ff76dd32de\") " Dec 01 08:59:35 crc kubenswrapper[4873]: I1201 08:59:35.253616 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c1fee8d-dc8f-4167-8d20-21ff76dd32de-kube-api-access-bkz7k" (OuterVolumeSpecName: "kube-api-access-bkz7k") pod "3c1fee8d-dc8f-4167-8d20-21ff76dd32de" (UID: "3c1fee8d-dc8f-4167-8d20-21ff76dd32de"). InnerVolumeSpecName "kube-api-access-bkz7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:59:35 crc kubenswrapper[4873]: I1201 08:59:35.294438 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c1fee8d-dc8f-4167-8d20-21ff76dd32de-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3c1fee8d-dc8f-4167-8d20-21ff76dd32de" (UID: "3c1fee8d-dc8f-4167-8d20-21ff76dd32de"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:59:35 crc kubenswrapper[4873]: I1201 08:59:35.299347 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c1fee8d-dc8f-4167-8d20-21ff76dd32de-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3c1fee8d-dc8f-4167-8d20-21ff76dd32de" (UID: "3c1fee8d-dc8f-4167-8d20-21ff76dd32de"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:59:35 crc kubenswrapper[4873]: I1201 08:59:35.306009 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c1fee8d-dc8f-4167-8d20-21ff76dd32de-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3c1fee8d-dc8f-4167-8d20-21ff76dd32de" (UID: "3c1fee8d-dc8f-4167-8d20-21ff76dd32de"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:59:35 crc kubenswrapper[4873]: I1201 08:59:35.307817 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c1fee8d-dc8f-4167-8d20-21ff76dd32de-config" (OuterVolumeSpecName: "config") pod "3c1fee8d-dc8f-4167-8d20-21ff76dd32de" (UID: "3c1fee8d-dc8f-4167-8d20-21ff76dd32de"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:59:35 crc kubenswrapper[4873]: I1201 08:59:35.357144 4873 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3c1fee8d-dc8f-4167-8d20-21ff76dd32de-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:35 crc kubenswrapper[4873]: I1201 08:59:35.357185 4873 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3c1fee8d-dc8f-4167-8d20-21ff76dd32de-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:35 crc kubenswrapper[4873]: I1201 08:59:35.357197 4873 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3c1fee8d-dc8f-4167-8d20-21ff76dd32de-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:35 crc kubenswrapper[4873]: I1201 08:59:35.357209 4873 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c1fee8d-dc8f-4167-8d20-21ff76dd32de-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:35 crc kubenswrapper[4873]: I1201 08:59:35.357220 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bkz7k\" (UniqueName: \"kubernetes.io/projected/3c1fee8d-dc8f-4167-8d20-21ff76dd32de-kube-api-access-bkz7k\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:35 crc kubenswrapper[4873]: I1201 08:59:35.765738 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-7xln4"] Dec 01 08:59:35 crc kubenswrapper[4873]: I1201 08:59:35.772942 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-7xln4"] Dec 01 08:59:35 crc kubenswrapper[4873]: I1201 08:59:35.854155 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54f9b7b8d9-t4rtz" event={"ID":"3c1fee8d-dc8f-4167-8d20-21ff76dd32de","Type":"ContainerDied","Data":"bb35b12d7a472bcbdc9544e9d227222eeea6c3188aafbbfcb038657f5128cbd5"} Dec 01 08:59:35 crc kubenswrapper[4873]: I1201 08:59:35.854297 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54f9b7b8d9-t4rtz" Dec 01 08:59:35 crc kubenswrapper[4873]: I1201 08:59:35.876561 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-kzt5x"] Dec 01 08:59:35 crc kubenswrapper[4873]: E1201 08:59:35.877144 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04cc97ba-b3a6-4906-87d1-b94c44858e86" containerName="init" Dec 01 08:59:35 crc kubenswrapper[4873]: I1201 08:59:35.877168 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="04cc97ba-b3a6-4906-87d1-b94c44858e86" containerName="init" Dec 01 08:59:35 crc kubenswrapper[4873]: E1201 08:59:35.877195 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ac6fb56-fa10-4936-b8b3-1204d14945f8" containerName="keystone-bootstrap" Dec 01 08:59:35 crc kubenswrapper[4873]: I1201 08:59:35.877207 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ac6fb56-fa10-4936-b8b3-1204d14945f8" containerName="keystone-bootstrap" Dec 01 08:59:35 crc kubenswrapper[4873]: E1201 08:59:35.877221 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c1fee8d-dc8f-4167-8d20-21ff76dd32de" containerName="dnsmasq-dns" Dec 01 08:59:35 crc kubenswrapper[4873]: I1201 08:59:35.877229 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c1fee8d-dc8f-4167-8d20-21ff76dd32de" containerName="dnsmasq-dns" Dec 01 08:59:35 crc kubenswrapper[4873]: E1201 08:59:35.877256 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c1fee8d-dc8f-4167-8d20-21ff76dd32de" containerName="init" Dec 01 08:59:35 crc kubenswrapper[4873]: I1201 08:59:35.877265 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c1fee8d-dc8f-4167-8d20-21ff76dd32de" containerName="init" Dec 01 08:59:35 crc kubenswrapper[4873]: I1201 08:59:35.877540 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ac6fb56-fa10-4936-b8b3-1204d14945f8" containerName="keystone-bootstrap" Dec 01 08:59:35 crc kubenswrapper[4873]: I1201 08:59:35.877560 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c1fee8d-dc8f-4167-8d20-21ff76dd32de" containerName="dnsmasq-dns" Dec 01 08:59:35 crc kubenswrapper[4873]: I1201 08:59:35.877570 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="04cc97ba-b3a6-4906-87d1-b94c44858e86" containerName="init" Dec 01 08:59:35 crc kubenswrapper[4873]: I1201 08:59:35.879919 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-kzt5x" Dec 01 08:59:35 crc kubenswrapper[4873]: I1201 08:59:35.886840 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 01 08:59:35 crc kubenswrapper[4873]: I1201 08:59:35.886882 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 01 08:59:35 crc kubenswrapper[4873]: I1201 08:59:35.886893 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-9jwfx" Dec 01 08:59:35 crc kubenswrapper[4873]: I1201 08:59:35.887056 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 01 08:59:35 crc kubenswrapper[4873]: I1201 08:59:35.887117 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 01 08:59:35 crc kubenswrapper[4873]: I1201 08:59:35.895194 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-kzt5x"] Dec 01 08:59:35 crc kubenswrapper[4873]: I1201 08:59:35.916442 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54f9b7b8d9-t4rtz"] Dec 01 08:59:35 crc kubenswrapper[4873]: I1201 08:59:35.934430 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-54f9b7b8d9-t4rtz"] Dec 01 08:59:36 crc kubenswrapper[4873]: I1201 08:59:36.070652 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a34b9de-9fde-4ff2-9881-1dcbea700c28-config-data\") pod \"keystone-bootstrap-kzt5x\" (UID: \"0a34b9de-9fde-4ff2-9881-1dcbea700c28\") " pod="openstack/keystone-bootstrap-kzt5x" Dec 01 08:59:36 crc kubenswrapper[4873]: I1201 08:59:36.070725 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0a34b9de-9fde-4ff2-9881-1dcbea700c28-fernet-keys\") pod \"keystone-bootstrap-kzt5x\" (UID: \"0a34b9de-9fde-4ff2-9881-1dcbea700c28\") " pod="openstack/keystone-bootstrap-kzt5x" Dec 01 08:59:36 crc kubenswrapper[4873]: I1201 08:59:36.070757 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0a34b9de-9fde-4ff2-9881-1dcbea700c28-credential-keys\") pod \"keystone-bootstrap-kzt5x\" (UID: \"0a34b9de-9fde-4ff2-9881-1dcbea700c28\") " pod="openstack/keystone-bootstrap-kzt5x" Dec 01 08:59:36 crc kubenswrapper[4873]: I1201 08:59:36.070834 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a34b9de-9fde-4ff2-9881-1dcbea700c28-scripts\") pod \"keystone-bootstrap-kzt5x\" (UID: \"0a34b9de-9fde-4ff2-9881-1dcbea700c28\") " pod="openstack/keystone-bootstrap-kzt5x" Dec 01 08:59:36 crc kubenswrapper[4873]: I1201 08:59:36.070853 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77dmf\" (UniqueName: \"kubernetes.io/projected/0a34b9de-9fde-4ff2-9881-1dcbea700c28-kube-api-access-77dmf\") pod \"keystone-bootstrap-kzt5x\" (UID: \"0a34b9de-9fde-4ff2-9881-1dcbea700c28\") " pod="openstack/keystone-bootstrap-kzt5x" Dec 01 08:59:36 crc kubenswrapper[4873]: I1201 08:59:36.071162 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a34b9de-9fde-4ff2-9881-1dcbea700c28-combined-ca-bundle\") pod \"keystone-bootstrap-kzt5x\" (UID: \"0a34b9de-9fde-4ff2-9881-1dcbea700c28\") " pod="openstack/keystone-bootstrap-kzt5x" Dec 01 08:59:36 crc kubenswrapper[4873]: I1201 08:59:36.173203 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a34b9de-9fde-4ff2-9881-1dcbea700c28-combined-ca-bundle\") pod \"keystone-bootstrap-kzt5x\" (UID: \"0a34b9de-9fde-4ff2-9881-1dcbea700c28\") " pod="openstack/keystone-bootstrap-kzt5x" Dec 01 08:59:36 crc kubenswrapper[4873]: I1201 08:59:36.173278 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a34b9de-9fde-4ff2-9881-1dcbea700c28-config-data\") pod \"keystone-bootstrap-kzt5x\" (UID: \"0a34b9de-9fde-4ff2-9881-1dcbea700c28\") " pod="openstack/keystone-bootstrap-kzt5x" Dec 01 08:59:36 crc kubenswrapper[4873]: I1201 08:59:36.173310 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0a34b9de-9fde-4ff2-9881-1dcbea700c28-fernet-keys\") pod \"keystone-bootstrap-kzt5x\" (UID: \"0a34b9de-9fde-4ff2-9881-1dcbea700c28\") " pod="openstack/keystone-bootstrap-kzt5x" Dec 01 08:59:36 crc kubenswrapper[4873]: I1201 08:59:36.173338 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0a34b9de-9fde-4ff2-9881-1dcbea700c28-credential-keys\") pod \"keystone-bootstrap-kzt5x\" (UID: \"0a34b9de-9fde-4ff2-9881-1dcbea700c28\") " pod="openstack/keystone-bootstrap-kzt5x" Dec 01 08:59:36 crc kubenswrapper[4873]: I1201 08:59:36.173425 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a34b9de-9fde-4ff2-9881-1dcbea700c28-scripts\") pod \"keystone-bootstrap-kzt5x\" (UID: \"0a34b9de-9fde-4ff2-9881-1dcbea700c28\") " pod="openstack/keystone-bootstrap-kzt5x" Dec 01 08:59:36 crc kubenswrapper[4873]: I1201 08:59:36.173458 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77dmf\" (UniqueName: \"kubernetes.io/projected/0a34b9de-9fde-4ff2-9881-1dcbea700c28-kube-api-access-77dmf\") pod \"keystone-bootstrap-kzt5x\" (UID: \"0a34b9de-9fde-4ff2-9881-1dcbea700c28\") " pod="openstack/keystone-bootstrap-kzt5x" Dec 01 08:59:36 crc kubenswrapper[4873]: I1201 08:59:36.180061 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a34b9de-9fde-4ff2-9881-1dcbea700c28-config-data\") pod \"keystone-bootstrap-kzt5x\" (UID: \"0a34b9de-9fde-4ff2-9881-1dcbea700c28\") " pod="openstack/keystone-bootstrap-kzt5x" Dec 01 08:59:36 crc kubenswrapper[4873]: I1201 08:59:36.181533 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0a34b9de-9fde-4ff2-9881-1dcbea700c28-fernet-keys\") pod \"keystone-bootstrap-kzt5x\" (UID: \"0a34b9de-9fde-4ff2-9881-1dcbea700c28\") " pod="openstack/keystone-bootstrap-kzt5x" Dec 01 08:59:36 crc kubenswrapper[4873]: I1201 08:59:36.182876 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a34b9de-9fde-4ff2-9881-1dcbea700c28-combined-ca-bundle\") pod \"keystone-bootstrap-kzt5x\" (UID: \"0a34b9de-9fde-4ff2-9881-1dcbea700c28\") " pod="openstack/keystone-bootstrap-kzt5x" Dec 01 08:59:36 crc kubenswrapper[4873]: I1201 08:59:36.194455 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a34b9de-9fde-4ff2-9881-1dcbea700c28-scripts\") pod \"keystone-bootstrap-kzt5x\" (UID: \"0a34b9de-9fde-4ff2-9881-1dcbea700c28\") " pod="openstack/keystone-bootstrap-kzt5x" Dec 01 08:59:36 crc kubenswrapper[4873]: I1201 08:59:36.195831 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0a34b9de-9fde-4ff2-9881-1dcbea700c28-credential-keys\") pod \"keystone-bootstrap-kzt5x\" (UID: \"0a34b9de-9fde-4ff2-9881-1dcbea700c28\") " pod="openstack/keystone-bootstrap-kzt5x" Dec 01 08:59:36 crc kubenswrapper[4873]: I1201 08:59:36.208721 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77dmf\" (UniqueName: \"kubernetes.io/projected/0a34b9de-9fde-4ff2-9881-1dcbea700c28-kube-api-access-77dmf\") pod \"keystone-bootstrap-kzt5x\" (UID: \"0a34b9de-9fde-4ff2-9881-1dcbea700c28\") " pod="openstack/keystone-bootstrap-kzt5x" Dec 01 08:59:36 crc kubenswrapper[4873]: I1201 08:59:36.213994 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-kzt5x" Dec 01 08:59:36 crc kubenswrapper[4873]: I1201 08:59:36.443583 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c1fee8d-dc8f-4167-8d20-21ff76dd32de" path="/var/lib/kubelet/pods/3c1fee8d-dc8f-4167-8d20-21ff76dd32de/volumes" Dec 01 08:59:36 crc kubenswrapper[4873]: I1201 08:59:36.444692 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ac6fb56-fa10-4936-b8b3-1204d14945f8" path="/var/lib/kubelet/pods/7ac6fb56-fa10-4936-b8b3-1204d14945f8/volumes" Dec 01 08:59:44 crc kubenswrapper[4873]: E1201 08:59:44.743148 4873 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Dec 01 08:59:44 crc kubenswrapper[4873]: E1201 08:59:44.744269 4873 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-h64g6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-9lvtd_openstack(c8c5adca-9738-46b1-9376-01cfb425d839): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 08:59:44 crc kubenswrapper[4873]: I1201 08:59:44.743791 4873 scope.go:117] "RemoveContainer" containerID="539f87537745220b07fb39faf639760032a7e74d7310a53693c1981fff1fc864" Dec 01 08:59:44 crc kubenswrapper[4873]: E1201 08:59:44.746370 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-9lvtd" podUID="c8c5adca-9738-46b1-9376-01cfb425d839" Dec 01 08:59:45 crc kubenswrapper[4873]: E1201 08:59:45.055979 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-9lvtd" podUID="c8c5adca-9738-46b1-9376-01cfb425d839" Dec 01 08:59:46 crc kubenswrapper[4873]: I1201 08:59:46.009133 4873 scope.go:117] "RemoveContainer" containerID="0a167f1c59cd2ef54a4dee50fa3d2e584b21635e5a7cfc13cf2bea48e1132b42" Dec 01 08:59:46 crc kubenswrapper[4873]: E1201 08:59:46.019403 4873 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 01 08:59:46 crc kubenswrapper[4873]: E1201 08:59:46.019927 4873 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rtl44,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-bmhcv_openstack(ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 08:59:46 crc kubenswrapper[4873]: E1201 08:59:46.022376 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-bmhcv" podUID="ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce" Dec 01 08:59:46 crc kubenswrapper[4873]: E1201 08:59:46.073724 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-bmhcv" podUID="ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce" Dec 01 08:59:46 crc kubenswrapper[4873]: I1201 08:59:46.481375 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-kzt5x"] Dec 01 08:59:46 crc kubenswrapper[4873]: W1201 08:59:46.487233 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0a34b9de_9fde_4ff2_9881_1dcbea700c28.slice/crio-6a054c124b087ef0ce976dca7ed64aa3e9604be8a0039ca56f50479e1de1e364 WatchSource:0}: Error finding container 6a054c124b087ef0ce976dca7ed64aa3e9604be8a0039ca56f50479e1de1e364: Status 404 returned error can't find the container with id 6a054c124b087ef0ce976dca7ed64aa3e9604be8a0039ca56f50479e1de1e364 Dec 01 08:59:47 crc kubenswrapper[4873]: I1201 08:59:47.089424 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" event={"ID":"fef7b114-0e07-402d-a37b-315c36011f4b","Type":"ContainerStarted","Data":"90066b4e910f0ee21a175275171e35716a5459a2c723385e352a837980db5011"} Dec 01 08:59:47 crc kubenswrapper[4873]: I1201 08:59:47.091611 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-kzt5x" event={"ID":"0a34b9de-9fde-4ff2-9881-1dcbea700c28","Type":"ContainerStarted","Data":"14a602dd9e3b3f009e5a5469016a9cbc0c58a301f0f38499ba6035bb682849a6"} Dec 01 08:59:47 crc kubenswrapper[4873]: I1201 08:59:47.091668 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-kzt5x" event={"ID":"0a34b9de-9fde-4ff2-9881-1dcbea700c28","Type":"ContainerStarted","Data":"6a054c124b087ef0ce976dca7ed64aa3e9604be8a0039ca56f50479e1de1e364"} Dec 01 08:59:47 crc kubenswrapper[4873]: I1201 08:59:47.097095 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-nmmcg" event={"ID":"b0351e0d-24dc-407a-ba64-a38b313cf17c","Type":"ContainerStarted","Data":"60ccf9b2d1e68f6660316983c445b3c311c62d81172fe847a0d3d879b5fa6c77"} Dec 01 08:59:47 crc kubenswrapper[4873]: I1201 08:59:47.100198 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e53689a0-5d92-4e62-bb63-3a56af762a59","Type":"ContainerStarted","Data":"31940bfd155d788c99f80b4389ebba5de7bb5078293bf263879bda05942db53b"} Dec 01 08:59:47 crc kubenswrapper[4873]: I1201 08:59:47.175895 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-nmmcg" podStartSLOduration=4.691461643 podStartE2EDuration="25.175868567s" podCreationTimestamp="2025-12-01 08:59:22 +0000 UTC" firstStartedPulling="2025-12-01 08:59:24.241871916 +0000 UTC m=+1140.143980455" lastFinishedPulling="2025-12-01 08:59:44.72627884 +0000 UTC m=+1160.628387379" observedRunningTime="2025-12-01 08:59:47.171788341 +0000 UTC m=+1163.073896870" watchObservedRunningTime="2025-12-01 08:59:47.175868567 +0000 UTC m=+1163.077977106" Dec 01 08:59:47 crc kubenswrapper[4873]: I1201 08:59:47.198910 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-kzt5x" podStartSLOduration=12.19888145 podStartE2EDuration="12.19888145s" podCreationTimestamp="2025-12-01 08:59:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:59:47.190509433 +0000 UTC m=+1163.092617972" watchObservedRunningTime="2025-12-01 08:59:47.19888145 +0000 UTC m=+1163.100989989" Dec 01 08:59:48 crc kubenswrapper[4873]: I1201 08:59:48.120420 4873 generic.go:334] "Generic (PLEG): container finished" podID="ec6a6ffa-2ab6-41c4-b594-17066d11654a" containerID="df48d3b6b18f16bbf985f991b069677905d37acec7b9825b5ae4eb31fc298878" exitCode=0 Dec 01 08:59:48 crc kubenswrapper[4873]: I1201 08:59:48.120537 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-5vlz5" event={"ID":"ec6a6ffa-2ab6-41c4-b594-17066d11654a","Type":"ContainerDied","Data":"df48d3b6b18f16bbf985f991b069677905d37acec7b9825b5ae4eb31fc298878"} Dec 01 08:59:49 crc kubenswrapper[4873]: I1201 08:59:49.139388 4873 generic.go:334] "Generic (PLEG): container finished" podID="b0351e0d-24dc-407a-ba64-a38b313cf17c" containerID="60ccf9b2d1e68f6660316983c445b3c311c62d81172fe847a0d3d879b5fa6c77" exitCode=0 Dec 01 08:59:49 crc kubenswrapper[4873]: I1201 08:59:49.139489 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-nmmcg" event={"ID":"b0351e0d-24dc-407a-ba64-a38b313cf17c","Type":"ContainerDied","Data":"60ccf9b2d1e68f6660316983c445b3c311c62d81172fe847a0d3d879b5fa6c77"} Dec 01 08:59:49 crc kubenswrapper[4873]: I1201 08:59:49.148063 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e53689a0-5d92-4e62-bb63-3a56af762a59","Type":"ContainerStarted","Data":"0b43d48616acc5de469072536f3377110552ad3499b67b694a963f6c3a326cbc"} Dec 01 08:59:49 crc kubenswrapper[4873]: I1201 08:59:49.616630 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-5vlz5" Dec 01 08:59:49 crc kubenswrapper[4873]: I1201 08:59:49.734712 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec6a6ffa-2ab6-41c4-b594-17066d11654a-combined-ca-bundle\") pod \"ec6a6ffa-2ab6-41c4-b594-17066d11654a\" (UID: \"ec6a6ffa-2ab6-41c4-b594-17066d11654a\") " Dec 01 08:59:49 crc kubenswrapper[4873]: I1201 08:59:49.734785 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ec6a6ffa-2ab6-41c4-b594-17066d11654a-config\") pod \"ec6a6ffa-2ab6-41c4-b594-17066d11654a\" (UID: \"ec6a6ffa-2ab6-41c4-b594-17066d11654a\") " Dec 01 08:59:49 crc kubenswrapper[4873]: I1201 08:59:49.734829 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qkmlc\" (UniqueName: \"kubernetes.io/projected/ec6a6ffa-2ab6-41c4-b594-17066d11654a-kube-api-access-qkmlc\") pod \"ec6a6ffa-2ab6-41c4-b594-17066d11654a\" (UID: \"ec6a6ffa-2ab6-41c4-b594-17066d11654a\") " Dec 01 08:59:49 crc kubenswrapper[4873]: I1201 08:59:49.754977 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec6a6ffa-2ab6-41c4-b594-17066d11654a-kube-api-access-qkmlc" (OuterVolumeSpecName: "kube-api-access-qkmlc") pod "ec6a6ffa-2ab6-41c4-b594-17066d11654a" (UID: "ec6a6ffa-2ab6-41c4-b594-17066d11654a"). InnerVolumeSpecName "kube-api-access-qkmlc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:59:49 crc kubenswrapper[4873]: I1201 08:59:49.763900 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec6a6ffa-2ab6-41c4-b594-17066d11654a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ec6a6ffa-2ab6-41c4-b594-17066d11654a" (UID: "ec6a6ffa-2ab6-41c4-b594-17066d11654a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:59:49 crc kubenswrapper[4873]: I1201 08:59:49.779799 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec6a6ffa-2ab6-41c4-b594-17066d11654a-config" (OuterVolumeSpecName: "config") pod "ec6a6ffa-2ab6-41c4-b594-17066d11654a" (UID: "ec6a6ffa-2ab6-41c4-b594-17066d11654a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:59:49 crc kubenswrapper[4873]: I1201 08:59:49.837577 4873 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec6a6ffa-2ab6-41c4-b594-17066d11654a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:49 crc kubenswrapper[4873]: I1201 08:59:49.837614 4873 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/ec6a6ffa-2ab6-41c4-b594-17066d11654a-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:49 crc kubenswrapper[4873]: I1201 08:59:49.837628 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qkmlc\" (UniqueName: \"kubernetes.io/projected/ec6a6ffa-2ab6-41c4-b594-17066d11654a-kube-api-access-qkmlc\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.162849 4873 generic.go:334] "Generic (PLEG): container finished" podID="0a34b9de-9fde-4ff2-9881-1dcbea700c28" containerID="14a602dd9e3b3f009e5a5469016a9cbc0c58a301f0f38499ba6035bb682849a6" exitCode=0 Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.163459 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-kzt5x" event={"ID":"0a34b9de-9fde-4ff2-9881-1dcbea700c28","Type":"ContainerDied","Data":"14a602dd9e3b3f009e5a5469016a9cbc0c58a301f0f38499ba6035bb682849a6"} Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.167250 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-5vlz5" Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.171474 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-5vlz5" event={"ID":"ec6a6ffa-2ab6-41c4-b594-17066d11654a","Type":"ContainerDied","Data":"3b3b15f9d1116acf24cf5f574e1c3e9cadd04a88ee84b3b58e3be1104f470fb4"} Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.171516 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b3b15f9d1116acf24cf5f574e1c3e9cadd04a88ee84b3b58e3be1104f470fb4" Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.424206 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7b946d459c-csc2t"] Dec 01 08:59:50 crc kubenswrapper[4873]: E1201 08:59:50.425241 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec6a6ffa-2ab6-41c4-b594-17066d11654a" containerName="neutron-db-sync" Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.425270 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec6a6ffa-2ab6-41c4-b594-17066d11654a" containerName="neutron-db-sync" Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.450858 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec6a6ffa-2ab6-41c4-b594-17066d11654a" containerName="neutron-db-sync" Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.457548 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b946d459c-csc2t" Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.544316 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b946d459c-csc2t"] Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.593140 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/21098264-17f0-46eb-8531-7df74404e5c7-dns-svc\") pod \"dnsmasq-dns-7b946d459c-csc2t\" (UID: \"21098264-17f0-46eb-8531-7df74404e5c7\") " pod="openstack/dnsmasq-dns-7b946d459c-csc2t" Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.593248 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7wpt\" (UniqueName: \"kubernetes.io/projected/21098264-17f0-46eb-8531-7df74404e5c7-kube-api-access-s7wpt\") pod \"dnsmasq-dns-7b946d459c-csc2t\" (UID: \"21098264-17f0-46eb-8531-7df74404e5c7\") " pod="openstack/dnsmasq-dns-7b946d459c-csc2t" Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.593294 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/21098264-17f0-46eb-8531-7df74404e5c7-ovsdbserver-sb\") pod \"dnsmasq-dns-7b946d459c-csc2t\" (UID: \"21098264-17f0-46eb-8531-7df74404e5c7\") " pod="openstack/dnsmasq-dns-7b946d459c-csc2t" Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.593325 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/21098264-17f0-46eb-8531-7df74404e5c7-ovsdbserver-nb\") pod \"dnsmasq-dns-7b946d459c-csc2t\" (UID: \"21098264-17f0-46eb-8531-7df74404e5c7\") " pod="openstack/dnsmasq-dns-7b946d459c-csc2t" Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.593359 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21098264-17f0-46eb-8531-7df74404e5c7-config\") pod \"dnsmasq-dns-7b946d459c-csc2t\" (UID: \"21098264-17f0-46eb-8531-7df74404e5c7\") " pod="openstack/dnsmasq-dns-7b946d459c-csc2t" Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.596982 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-56d597fbb8-m5cgw"] Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.598772 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-56d597fbb8-m5cgw" Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.604646 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.604934 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.605133 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.607347 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-92gc9" Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.615665 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-56d597fbb8-m5cgw"] Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.695377 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7wpt\" (UniqueName: \"kubernetes.io/projected/21098264-17f0-46eb-8531-7df74404e5c7-kube-api-access-s7wpt\") pod \"dnsmasq-dns-7b946d459c-csc2t\" (UID: \"21098264-17f0-46eb-8531-7df74404e5c7\") " pod="openstack/dnsmasq-dns-7b946d459c-csc2t" Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.695464 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/21098264-17f0-46eb-8531-7df74404e5c7-ovsdbserver-sb\") pod \"dnsmasq-dns-7b946d459c-csc2t\" (UID: \"21098264-17f0-46eb-8531-7df74404e5c7\") " pod="openstack/dnsmasq-dns-7b946d459c-csc2t" Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.695510 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/21098264-17f0-46eb-8531-7df74404e5c7-ovsdbserver-nb\") pod \"dnsmasq-dns-7b946d459c-csc2t\" (UID: \"21098264-17f0-46eb-8531-7df74404e5c7\") " pod="openstack/dnsmasq-dns-7b946d459c-csc2t" Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.695551 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21098264-17f0-46eb-8531-7df74404e5c7-config\") pod \"dnsmasq-dns-7b946d459c-csc2t\" (UID: \"21098264-17f0-46eb-8531-7df74404e5c7\") " pod="openstack/dnsmasq-dns-7b946d459c-csc2t" Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.695630 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/21098264-17f0-46eb-8531-7df74404e5c7-dns-svc\") pod \"dnsmasq-dns-7b946d459c-csc2t\" (UID: \"21098264-17f0-46eb-8531-7df74404e5c7\") " pod="openstack/dnsmasq-dns-7b946d459c-csc2t" Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.696814 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/21098264-17f0-46eb-8531-7df74404e5c7-dns-svc\") pod \"dnsmasq-dns-7b946d459c-csc2t\" (UID: \"21098264-17f0-46eb-8531-7df74404e5c7\") " pod="openstack/dnsmasq-dns-7b946d459c-csc2t" Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.698888 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/21098264-17f0-46eb-8531-7df74404e5c7-ovsdbserver-nb\") pod \"dnsmasq-dns-7b946d459c-csc2t\" (UID: \"21098264-17f0-46eb-8531-7df74404e5c7\") " pod="openstack/dnsmasq-dns-7b946d459c-csc2t" Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.699143 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/21098264-17f0-46eb-8531-7df74404e5c7-ovsdbserver-sb\") pod \"dnsmasq-dns-7b946d459c-csc2t\" (UID: \"21098264-17f0-46eb-8531-7df74404e5c7\") " pod="openstack/dnsmasq-dns-7b946d459c-csc2t" Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.699367 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21098264-17f0-46eb-8531-7df74404e5c7-config\") pod \"dnsmasq-dns-7b946d459c-csc2t\" (UID: \"21098264-17f0-46eb-8531-7df74404e5c7\") " pod="openstack/dnsmasq-dns-7b946d459c-csc2t" Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.743536 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7wpt\" (UniqueName: \"kubernetes.io/projected/21098264-17f0-46eb-8531-7df74404e5c7-kube-api-access-s7wpt\") pod \"dnsmasq-dns-7b946d459c-csc2t\" (UID: \"21098264-17f0-46eb-8531-7df74404e5c7\") " pod="openstack/dnsmasq-dns-7b946d459c-csc2t" Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.797952 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrrw2\" (UniqueName: \"kubernetes.io/projected/1a53851f-14c9-4577-bb4e-c3a7b6b8c141-kube-api-access-vrrw2\") pod \"neutron-56d597fbb8-m5cgw\" (UID: \"1a53851f-14c9-4577-bb4e-c3a7b6b8c141\") " pod="openstack/neutron-56d597fbb8-m5cgw" Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.798507 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1a53851f-14c9-4577-bb4e-c3a7b6b8c141-httpd-config\") pod \"neutron-56d597fbb8-m5cgw\" (UID: \"1a53851f-14c9-4577-bb4e-c3a7b6b8c141\") " pod="openstack/neutron-56d597fbb8-m5cgw" Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.798556 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1a53851f-14c9-4577-bb4e-c3a7b6b8c141-config\") pod \"neutron-56d597fbb8-m5cgw\" (UID: \"1a53851f-14c9-4577-bb4e-c3a7b6b8c141\") " pod="openstack/neutron-56d597fbb8-m5cgw" Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.798587 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a53851f-14c9-4577-bb4e-c3a7b6b8c141-combined-ca-bundle\") pod \"neutron-56d597fbb8-m5cgw\" (UID: \"1a53851f-14c9-4577-bb4e-c3a7b6b8c141\") " pod="openstack/neutron-56d597fbb8-m5cgw" Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.798656 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a53851f-14c9-4577-bb4e-c3a7b6b8c141-ovndb-tls-certs\") pod \"neutron-56d597fbb8-m5cgw\" (UID: \"1a53851f-14c9-4577-bb4e-c3a7b6b8c141\") " pod="openstack/neutron-56d597fbb8-m5cgw" Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.850880 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b946d459c-csc2t" Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.900840 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrrw2\" (UniqueName: \"kubernetes.io/projected/1a53851f-14c9-4577-bb4e-c3a7b6b8c141-kube-api-access-vrrw2\") pod \"neutron-56d597fbb8-m5cgw\" (UID: \"1a53851f-14c9-4577-bb4e-c3a7b6b8c141\") " pod="openstack/neutron-56d597fbb8-m5cgw" Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.900936 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1a53851f-14c9-4577-bb4e-c3a7b6b8c141-httpd-config\") pod \"neutron-56d597fbb8-m5cgw\" (UID: \"1a53851f-14c9-4577-bb4e-c3a7b6b8c141\") " pod="openstack/neutron-56d597fbb8-m5cgw" Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.900985 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1a53851f-14c9-4577-bb4e-c3a7b6b8c141-config\") pod \"neutron-56d597fbb8-m5cgw\" (UID: \"1a53851f-14c9-4577-bb4e-c3a7b6b8c141\") " pod="openstack/neutron-56d597fbb8-m5cgw" Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.901018 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a53851f-14c9-4577-bb4e-c3a7b6b8c141-combined-ca-bundle\") pod \"neutron-56d597fbb8-m5cgw\" (UID: \"1a53851f-14c9-4577-bb4e-c3a7b6b8c141\") " pod="openstack/neutron-56d597fbb8-m5cgw" Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.901155 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a53851f-14c9-4577-bb4e-c3a7b6b8c141-ovndb-tls-certs\") pod \"neutron-56d597fbb8-m5cgw\" (UID: \"1a53851f-14c9-4577-bb4e-c3a7b6b8c141\") " pod="openstack/neutron-56d597fbb8-m5cgw" Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.909322 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1a53851f-14c9-4577-bb4e-c3a7b6b8c141-httpd-config\") pod \"neutron-56d597fbb8-m5cgw\" (UID: \"1a53851f-14c9-4577-bb4e-c3a7b6b8c141\") " pod="openstack/neutron-56d597fbb8-m5cgw" Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.917098 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a53851f-14c9-4577-bb4e-c3a7b6b8c141-combined-ca-bundle\") pod \"neutron-56d597fbb8-m5cgw\" (UID: \"1a53851f-14c9-4577-bb4e-c3a7b6b8c141\") " pod="openstack/neutron-56d597fbb8-m5cgw" Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.927227 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/1a53851f-14c9-4577-bb4e-c3a7b6b8c141-config\") pod \"neutron-56d597fbb8-m5cgw\" (UID: \"1a53851f-14c9-4577-bb4e-c3a7b6b8c141\") " pod="openstack/neutron-56d597fbb8-m5cgw" Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.927246 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a53851f-14c9-4577-bb4e-c3a7b6b8c141-ovndb-tls-certs\") pod \"neutron-56d597fbb8-m5cgw\" (UID: \"1a53851f-14c9-4577-bb4e-c3a7b6b8c141\") " pod="openstack/neutron-56d597fbb8-m5cgw" Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.928803 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrrw2\" (UniqueName: \"kubernetes.io/projected/1a53851f-14c9-4577-bb4e-c3a7b6b8c141-kube-api-access-vrrw2\") pod \"neutron-56d597fbb8-m5cgw\" (UID: \"1a53851f-14c9-4577-bb4e-c3a7b6b8c141\") " pod="openstack/neutron-56d597fbb8-m5cgw" Dec 01 08:59:50 crc kubenswrapper[4873]: I1201 08:59:50.936963 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-56d597fbb8-m5cgw" Dec 01 08:59:53 crc kubenswrapper[4873]: I1201 08:59:53.344476 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-f99dddf57-sr27n"] Dec 01 08:59:53 crc kubenswrapper[4873]: I1201 08:59:53.348557 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f99dddf57-sr27n" Dec 01 08:59:53 crc kubenswrapper[4873]: I1201 08:59:53.351269 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 01 08:59:53 crc kubenswrapper[4873]: I1201 08:59:53.353380 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 01 08:59:53 crc kubenswrapper[4873]: I1201 08:59:53.371589 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-f99dddf57-sr27n"] Dec 01 08:59:53 crc kubenswrapper[4873]: I1201 08:59:53.541959 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b48bbdb-ef25-4a25-a79f-d01eca3d63df-public-tls-certs\") pod \"neutron-f99dddf57-sr27n\" (UID: \"7b48bbdb-ef25-4a25-a79f-d01eca3d63df\") " pod="openstack/neutron-f99dddf57-sr27n" Dec 01 08:59:53 crc kubenswrapper[4873]: I1201 08:59:53.542104 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b48bbdb-ef25-4a25-a79f-d01eca3d63df-ovndb-tls-certs\") pod \"neutron-f99dddf57-sr27n\" (UID: \"7b48bbdb-ef25-4a25-a79f-d01eca3d63df\") " pod="openstack/neutron-f99dddf57-sr27n" Dec 01 08:59:53 crc kubenswrapper[4873]: I1201 08:59:53.542137 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b48bbdb-ef25-4a25-a79f-d01eca3d63df-combined-ca-bundle\") pod \"neutron-f99dddf57-sr27n\" (UID: \"7b48bbdb-ef25-4a25-a79f-d01eca3d63df\") " pod="openstack/neutron-f99dddf57-sr27n" Dec 01 08:59:53 crc kubenswrapper[4873]: I1201 08:59:53.542165 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7b48bbdb-ef25-4a25-a79f-d01eca3d63df-config\") pod \"neutron-f99dddf57-sr27n\" (UID: \"7b48bbdb-ef25-4a25-a79f-d01eca3d63df\") " pod="openstack/neutron-f99dddf57-sr27n" Dec 01 08:59:53 crc kubenswrapper[4873]: I1201 08:59:53.542347 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7b48bbdb-ef25-4a25-a79f-d01eca3d63df-httpd-config\") pod \"neutron-f99dddf57-sr27n\" (UID: \"7b48bbdb-ef25-4a25-a79f-d01eca3d63df\") " pod="openstack/neutron-f99dddf57-sr27n" Dec 01 08:59:53 crc kubenswrapper[4873]: I1201 08:59:53.542380 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b48bbdb-ef25-4a25-a79f-d01eca3d63df-internal-tls-certs\") pod \"neutron-f99dddf57-sr27n\" (UID: \"7b48bbdb-ef25-4a25-a79f-d01eca3d63df\") " pod="openstack/neutron-f99dddf57-sr27n" Dec 01 08:59:53 crc kubenswrapper[4873]: I1201 08:59:53.542412 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gjhb\" (UniqueName: \"kubernetes.io/projected/7b48bbdb-ef25-4a25-a79f-d01eca3d63df-kube-api-access-8gjhb\") pod \"neutron-f99dddf57-sr27n\" (UID: \"7b48bbdb-ef25-4a25-a79f-d01eca3d63df\") " pod="openstack/neutron-f99dddf57-sr27n" Dec 01 08:59:53 crc kubenswrapper[4873]: I1201 08:59:53.645551 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b48bbdb-ef25-4a25-a79f-d01eca3d63df-public-tls-certs\") pod \"neutron-f99dddf57-sr27n\" (UID: \"7b48bbdb-ef25-4a25-a79f-d01eca3d63df\") " pod="openstack/neutron-f99dddf57-sr27n" Dec 01 08:59:53 crc kubenswrapper[4873]: I1201 08:59:53.645672 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b48bbdb-ef25-4a25-a79f-d01eca3d63df-ovndb-tls-certs\") pod \"neutron-f99dddf57-sr27n\" (UID: \"7b48bbdb-ef25-4a25-a79f-d01eca3d63df\") " pod="openstack/neutron-f99dddf57-sr27n" Dec 01 08:59:53 crc kubenswrapper[4873]: I1201 08:59:53.645699 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b48bbdb-ef25-4a25-a79f-d01eca3d63df-combined-ca-bundle\") pod \"neutron-f99dddf57-sr27n\" (UID: \"7b48bbdb-ef25-4a25-a79f-d01eca3d63df\") " pod="openstack/neutron-f99dddf57-sr27n" Dec 01 08:59:53 crc kubenswrapper[4873]: I1201 08:59:53.645744 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7b48bbdb-ef25-4a25-a79f-d01eca3d63df-config\") pod \"neutron-f99dddf57-sr27n\" (UID: \"7b48bbdb-ef25-4a25-a79f-d01eca3d63df\") " pod="openstack/neutron-f99dddf57-sr27n" Dec 01 08:59:53 crc kubenswrapper[4873]: I1201 08:59:53.645785 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7b48bbdb-ef25-4a25-a79f-d01eca3d63df-httpd-config\") pod \"neutron-f99dddf57-sr27n\" (UID: \"7b48bbdb-ef25-4a25-a79f-d01eca3d63df\") " pod="openstack/neutron-f99dddf57-sr27n" Dec 01 08:59:53 crc kubenswrapper[4873]: I1201 08:59:53.645825 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b48bbdb-ef25-4a25-a79f-d01eca3d63df-internal-tls-certs\") pod \"neutron-f99dddf57-sr27n\" (UID: \"7b48bbdb-ef25-4a25-a79f-d01eca3d63df\") " pod="openstack/neutron-f99dddf57-sr27n" Dec 01 08:59:53 crc kubenswrapper[4873]: I1201 08:59:53.645883 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gjhb\" (UniqueName: \"kubernetes.io/projected/7b48bbdb-ef25-4a25-a79f-d01eca3d63df-kube-api-access-8gjhb\") pod \"neutron-f99dddf57-sr27n\" (UID: \"7b48bbdb-ef25-4a25-a79f-d01eca3d63df\") " pod="openstack/neutron-f99dddf57-sr27n" Dec 01 08:59:53 crc kubenswrapper[4873]: I1201 08:59:53.655817 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b48bbdb-ef25-4a25-a79f-d01eca3d63df-ovndb-tls-certs\") pod \"neutron-f99dddf57-sr27n\" (UID: \"7b48bbdb-ef25-4a25-a79f-d01eca3d63df\") " pod="openstack/neutron-f99dddf57-sr27n" Dec 01 08:59:53 crc kubenswrapper[4873]: I1201 08:59:53.656452 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b48bbdb-ef25-4a25-a79f-d01eca3d63df-public-tls-certs\") pod \"neutron-f99dddf57-sr27n\" (UID: \"7b48bbdb-ef25-4a25-a79f-d01eca3d63df\") " pod="openstack/neutron-f99dddf57-sr27n" Dec 01 08:59:53 crc kubenswrapper[4873]: I1201 08:59:53.657322 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/7b48bbdb-ef25-4a25-a79f-d01eca3d63df-config\") pod \"neutron-f99dddf57-sr27n\" (UID: \"7b48bbdb-ef25-4a25-a79f-d01eca3d63df\") " pod="openstack/neutron-f99dddf57-sr27n" Dec 01 08:59:53 crc kubenswrapper[4873]: I1201 08:59:53.658805 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b48bbdb-ef25-4a25-a79f-d01eca3d63df-internal-tls-certs\") pod \"neutron-f99dddf57-sr27n\" (UID: \"7b48bbdb-ef25-4a25-a79f-d01eca3d63df\") " pod="openstack/neutron-f99dddf57-sr27n" Dec 01 08:59:53 crc kubenswrapper[4873]: I1201 08:59:53.659722 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7b48bbdb-ef25-4a25-a79f-d01eca3d63df-httpd-config\") pod \"neutron-f99dddf57-sr27n\" (UID: \"7b48bbdb-ef25-4a25-a79f-d01eca3d63df\") " pod="openstack/neutron-f99dddf57-sr27n" Dec 01 08:59:53 crc kubenswrapper[4873]: I1201 08:59:53.660041 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b48bbdb-ef25-4a25-a79f-d01eca3d63df-combined-ca-bundle\") pod \"neutron-f99dddf57-sr27n\" (UID: \"7b48bbdb-ef25-4a25-a79f-d01eca3d63df\") " pod="openstack/neutron-f99dddf57-sr27n" Dec 01 08:59:53 crc kubenswrapper[4873]: I1201 08:59:53.666462 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gjhb\" (UniqueName: \"kubernetes.io/projected/7b48bbdb-ef25-4a25-a79f-d01eca3d63df-kube-api-access-8gjhb\") pod \"neutron-f99dddf57-sr27n\" (UID: \"7b48bbdb-ef25-4a25-a79f-d01eca3d63df\") " pod="openstack/neutron-f99dddf57-sr27n" Dec 01 08:59:53 crc kubenswrapper[4873]: I1201 08:59:53.696605 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f99dddf57-sr27n" Dec 01 08:59:54 crc kubenswrapper[4873]: I1201 08:59:54.513071 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-nmmcg" Dec 01 08:59:54 crc kubenswrapper[4873]: I1201 08:59:54.519582 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-kzt5x" Dec 01 08:59:54 crc kubenswrapper[4873]: I1201 08:59:54.592984 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b0351e0d-24dc-407a-ba64-a38b313cf17c-scripts\") pod \"b0351e0d-24dc-407a-ba64-a38b313cf17c\" (UID: \"b0351e0d-24dc-407a-ba64-a38b313cf17c\") " Dec 01 08:59:54 crc kubenswrapper[4873]: I1201 08:59:54.593589 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0351e0d-24dc-407a-ba64-a38b313cf17c-logs\") pod \"b0351e0d-24dc-407a-ba64-a38b313cf17c\" (UID: \"b0351e0d-24dc-407a-ba64-a38b313cf17c\") " Dec 01 08:59:54 crc kubenswrapper[4873]: I1201 08:59:54.593630 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0351e0d-24dc-407a-ba64-a38b313cf17c-combined-ca-bundle\") pod \"b0351e0d-24dc-407a-ba64-a38b313cf17c\" (UID: \"b0351e0d-24dc-407a-ba64-a38b313cf17c\") " Dec 01 08:59:54 crc kubenswrapper[4873]: I1201 08:59:54.593691 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0a34b9de-9fde-4ff2-9881-1dcbea700c28-credential-keys\") pod \"0a34b9de-9fde-4ff2-9881-1dcbea700c28\" (UID: \"0a34b9de-9fde-4ff2-9881-1dcbea700c28\") " Dec 01 08:59:54 crc kubenswrapper[4873]: I1201 08:59:54.593745 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a34b9de-9fde-4ff2-9881-1dcbea700c28-config-data\") pod \"0a34b9de-9fde-4ff2-9881-1dcbea700c28\" (UID: \"0a34b9de-9fde-4ff2-9881-1dcbea700c28\") " Dec 01 08:59:54 crc kubenswrapper[4873]: I1201 08:59:54.593807 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0a34b9de-9fde-4ff2-9881-1dcbea700c28-fernet-keys\") pod \"0a34b9de-9fde-4ff2-9881-1dcbea700c28\" (UID: \"0a34b9de-9fde-4ff2-9881-1dcbea700c28\") " Dec 01 08:59:54 crc kubenswrapper[4873]: I1201 08:59:54.593853 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-77dmf\" (UniqueName: \"kubernetes.io/projected/0a34b9de-9fde-4ff2-9881-1dcbea700c28-kube-api-access-77dmf\") pod \"0a34b9de-9fde-4ff2-9881-1dcbea700c28\" (UID: \"0a34b9de-9fde-4ff2-9881-1dcbea700c28\") " Dec 01 08:59:54 crc kubenswrapper[4873]: I1201 08:59:54.594128 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xjqdf\" (UniqueName: \"kubernetes.io/projected/b0351e0d-24dc-407a-ba64-a38b313cf17c-kube-api-access-xjqdf\") pod \"b0351e0d-24dc-407a-ba64-a38b313cf17c\" (UID: \"b0351e0d-24dc-407a-ba64-a38b313cf17c\") " Dec 01 08:59:54 crc kubenswrapper[4873]: I1201 08:59:54.594188 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a34b9de-9fde-4ff2-9881-1dcbea700c28-scripts\") pod \"0a34b9de-9fde-4ff2-9881-1dcbea700c28\" (UID: \"0a34b9de-9fde-4ff2-9881-1dcbea700c28\") " Dec 01 08:59:54 crc kubenswrapper[4873]: I1201 08:59:54.595482 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0351e0d-24dc-407a-ba64-a38b313cf17c-logs" (OuterVolumeSpecName: "logs") pod "b0351e0d-24dc-407a-ba64-a38b313cf17c" (UID: "b0351e0d-24dc-407a-ba64-a38b313cf17c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:59:54 crc kubenswrapper[4873]: I1201 08:59:54.597075 4873 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0351e0d-24dc-407a-ba64-a38b313cf17c-logs\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:54 crc kubenswrapper[4873]: I1201 08:59:54.598771 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a34b9de-9fde-4ff2-9881-1dcbea700c28-kube-api-access-77dmf" (OuterVolumeSpecName: "kube-api-access-77dmf") pod "0a34b9de-9fde-4ff2-9881-1dcbea700c28" (UID: "0a34b9de-9fde-4ff2-9881-1dcbea700c28"). InnerVolumeSpecName "kube-api-access-77dmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:59:54 crc kubenswrapper[4873]: I1201 08:59:54.601539 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0351e0d-24dc-407a-ba64-a38b313cf17c-scripts" (OuterVolumeSpecName: "scripts") pod "b0351e0d-24dc-407a-ba64-a38b313cf17c" (UID: "b0351e0d-24dc-407a-ba64-a38b313cf17c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:59:54 crc kubenswrapper[4873]: I1201 08:59:54.602303 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a34b9de-9fde-4ff2-9881-1dcbea700c28-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "0a34b9de-9fde-4ff2-9881-1dcbea700c28" (UID: "0a34b9de-9fde-4ff2-9881-1dcbea700c28"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:59:54 crc kubenswrapper[4873]: I1201 08:59:54.603894 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0351e0d-24dc-407a-ba64-a38b313cf17c-kube-api-access-xjqdf" (OuterVolumeSpecName: "kube-api-access-xjqdf") pod "b0351e0d-24dc-407a-ba64-a38b313cf17c" (UID: "b0351e0d-24dc-407a-ba64-a38b313cf17c"). InnerVolumeSpecName "kube-api-access-xjqdf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:59:54 crc kubenswrapper[4873]: I1201 08:59:54.606299 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a34b9de-9fde-4ff2-9881-1dcbea700c28-scripts" (OuterVolumeSpecName: "scripts") pod "0a34b9de-9fde-4ff2-9881-1dcbea700c28" (UID: "0a34b9de-9fde-4ff2-9881-1dcbea700c28"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:59:54 crc kubenswrapper[4873]: I1201 08:59:54.626494 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a34b9de-9fde-4ff2-9881-1dcbea700c28-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "0a34b9de-9fde-4ff2-9881-1dcbea700c28" (UID: "0a34b9de-9fde-4ff2-9881-1dcbea700c28"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:59:54 crc kubenswrapper[4873]: I1201 08:59:54.663685 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a34b9de-9fde-4ff2-9881-1dcbea700c28-config-data" (OuterVolumeSpecName: "config-data") pod "0a34b9de-9fde-4ff2-9881-1dcbea700c28" (UID: "0a34b9de-9fde-4ff2-9881-1dcbea700c28"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:59:54 crc kubenswrapper[4873]: I1201 08:59:54.683890 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0351e0d-24dc-407a-ba64-a38b313cf17c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b0351e0d-24dc-407a-ba64-a38b313cf17c" (UID: "b0351e0d-24dc-407a-ba64-a38b313cf17c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:59:54 crc kubenswrapper[4873]: I1201 08:59:54.697457 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0351e0d-24dc-407a-ba64-a38b313cf17c-config-data\") pod \"b0351e0d-24dc-407a-ba64-a38b313cf17c\" (UID: \"b0351e0d-24dc-407a-ba64-a38b313cf17c\") " Dec 01 08:59:54 crc kubenswrapper[4873]: I1201 08:59:54.697561 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a34b9de-9fde-4ff2-9881-1dcbea700c28-combined-ca-bundle\") pod \"0a34b9de-9fde-4ff2-9881-1dcbea700c28\" (UID: \"0a34b9de-9fde-4ff2-9881-1dcbea700c28\") " Dec 01 08:59:54 crc kubenswrapper[4873]: I1201 08:59:54.698025 4873 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0351e0d-24dc-407a-ba64-a38b313cf17c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:54 crc kubenswrapper[4873]: I1201 08:59:54.698048 4873 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0a34b9de-9fde-4ff2-9881-1dcbea700c28-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:54 crc kubenswrapper[4873]: I1201 08:59:54.698063 4873 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a34b9de-9fde-4ff2-9881-1dcbea700c28-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:54 crc kubenswrapper[4873]: I1201 08:59:54.698102 4873 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0a34b9de-9fde-4ff2-9881-1dcbea700c28-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:54 crc kubenswrapper[4873]: I1201 08:59:54.698113 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-77dmf\" (UniqueName: \"kubernetes.io/projected/0a34b9de-9fde-4ff2-9881-1dcbea700c28-kube-api-access-77dmf\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:54 crc kubenswrapper[4873]: I1201 08:59:54.698125 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xjqdf\" (UniqueName: \"kubernetes.io/projected/b0351e0d-24dc-407a-ba64-a38b313cf17c-kube-api-access-xjqdf\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:54 crc kubenswrapper[4873]: I1201 08:59:54.698134 4873 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a34b9de-9fde-4ff2-9881-1dcbea700c28-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:54 crc kubenswrapper[4873]: I1201 08:59:54.698143 4873 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b0351e0d-24dc-407a-ba64-a38b313cf17c-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:54 crc kubenswrapper[4873]: I1201 08:59:54.730171 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a34b9de-9fde-4ff2-9881-1dcbea700c28-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0a34b9de-9fde-4ff2-9881-1dcbea700c28" (UID: "0a34b9de-9fde-4ff2-9881-1dcbea700c28"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:59:54 crc kubenswrapper[4873]: I1201 08:59:54.735009 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0351e0d-24dc-407a-ba64-a38b313cf17c-config-data" (OuterVolumeSpecName: "config-data") pod "b0351e0d-24dc-407a-ba64-a38b313cf17c" (UID: "b0351e0d-24dc-407a-ba64-a38b313cf17c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:59:54 crc kubenswrapper[4873]: I1201 08:59:54.800757 4873 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0351e0d-24dc-407a-ba64-a38b313cf17c-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:54 crc kubenswrapper[4873]: I1201 08:59:54.800800 4873 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a34b9de-9fde-4ff2-9881-1dcbea700c28-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.219117 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-nmmcg" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.219122 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-nmmcg" event={"ID":"b0351e0d-24dc-407a-ba64-a38b313cf17c","Type":"ContainerDied","Data":"16815ff095924e4ac372b522bcd3a5cbc6d13aec2fd2cc6d7fa125c7dc4e9b74"} Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.219665 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="16815ff095924e4ac372b522bcd3a5cbc6d13aec2fd2cc6d7fa125c7dc4e9b74" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.223823 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e53689a0-5d92-4e62-bb63-3a56af762a59","Type":"ContainerStarted","Data":"b33b0f231101926f73eefa68214880e4c14df8323cb5ce1c1750a3ab4b4e019d"} Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.424871 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-kzt5x" event={"ID":"0a34b9de-9fde-4ff2-9881-1dcbea700c28","Type":"ContainerDied","Data":"6a054c124b087ef0ce976dca7ed64aa3e9604be8a0039ca56f50479e1de1e364"} Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.424965 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6a054c124b087ef0ce976dca7ed64aa3e9604be8a0039ca56f50479e1de1e364" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.425073 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-kzt5x" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.434225 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b946d459c-csc2t"] Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.470734 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-56d597fbb8-m5cgw"] Dec 01 08:59:55 crc kubenswrapper[4873]: W1201 08:59:55.505507 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a53851f_14c9_4577_bb4e_c3a7b6b8c141.slice/crio-dac62a9725f74be0188bcea687e04bdd8d000362222db48caba4d305358af664 WatchSource:0}: Error finding container dac62a9725f74be0188bcea687e04bdd8d000362222db48caba4d305358af664: Status 404 returned error can't find the container with id dac62a9725f74be0188bcea687e04bdd8d000362222db48caba4d305358af664 Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.541802 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-f99dddf57-sr27n"] Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.696892 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-58576c9895-sk9tb"] Dec 01 08:59:55 crc kubenswrapper[4873]: E1201 08:59:55.700263 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a34b9de-9fde-4ff2-9881-1dcbea700c28" containerName="keystone-bootstrap" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.700514 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a34b9de-9fde-4ff2-9881-1dcbea700c28" containerName="keystone-bootstrap" Dec 01 08:59:55 crc kubenswrapper[4873]: E1201 08:59:55.700622 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0351e0d-24dc-407a-ba64-a38b313cf17c" containerName="placement-db-sync" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.700699 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0351e0d-24dc-407a-ba64-a38b313cf17c" containerName="placement-db-sync" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.726246 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0351e0d-24dc-407a-ba64-a38b313cf17c" containerName="placement-db-sync" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.726304 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a34b9de-9fde-4ff2-9881-1dcbea700c28" containerName="keystone-bootstrap" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.726865 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-58576c9895-sk9tb"] Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.726951 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-58576c9895-sk9tb" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.734916 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.736704 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0857cf38-58dd-416b-92ca-6b98800ff512-internal-tls-certs\") pod \"keystone-58576c9895-sk9tb\" (UID: \"0857cf38-58dd-416b-92ca-6b98800ff512\") " pod="openstack/keystone-58576c9895-sk9tb" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.736942 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0857cf38-58dd-416b-92ca-6b98800ff512-combined-ca-bundle\") pod \"keystone-58576c9895-sk9tb\" (UID: \"0857cf38-58dd-416b-92ca-6b98800ff512\") " pod="openstack/keystone-58576c9895-sk9tb" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.737082 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0857cf38-58dd-416b-92ca-6b98800ff512-scripts\") pod \"keystone-58576c9895-sk9tb\" (UID: \"0857cf38-58dd-416b-92ca-6b98800ff512\") " pod="openstack/keystone-58576c9895-sk9tb" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.737230 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfrgb\" (UniqueName: \"kubernetes.io/projected/0857cf38-58dd-416b-92ca-6b98800ff512-kube-api-access-nfrgb\") pod \"keystone-58576c9895-sk9tb\" (UID: \"0857cf38-58dd-416b-92ca-6b98800ff512\") " pod="openstack/keystone-58576c9895-sk9tb" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.737406 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0857cf38-58dd-416b-92ca-6b98800ff512-fernet-keys\") pod \"keystone-58576c9895-sk9tb\" (UID: \"0857cf38-58dd-416b-92ca-6b98800ff512\") " pod="openstack/keystone-58576c9895-sk9tb" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.737557 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0857cf38-58dd-416b-92ca-6b98800ff512-credential-keys\") pod \"keystone-58576c9895-sk9tb\" (UID: \"0857cf38-58dd-416b-92ca-6b98800ff512\") " pod="openstack/keystone-58576c9895-sk9tb" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.737711 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0857cf38-58dd-416b-92ca-6b98800ff512-config-data\") pod \"keystone-58576c9895-sk9tb\" (UID: \"0857cf38-58dd-416b-92ca-6b98800ff512\") " pod="openstack/keystone-58576c9895-sk9tb" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.737819 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0857cf38-58dd-416b-92ca-6b98800ff512-public-tls-certs\") pod \"keystone-58576c9895-sk9tb\" (UID: \"0857cf38-58dd-416b-92ca-6b98800ff512\") " pod="openstack/keystone-58576c9895-sk9tb" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.740652 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-9jwfx" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.740828 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.740937 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.741073 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.741670 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.777317 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-7d689bd78d-s9rtj"] Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.779126 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7d689bd78d-s9rtj" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.795942 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.796219 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.796329 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.796429 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.797222 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-fgj2w" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.801110 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7d689bd78d-s9rtj"] Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.844279 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6615f806-2d35-4500-a746-3fab7fbfde88-logs\") pod \"placement-7d689bd78d-s9rtj\" (UID: \"6615f806-2d35-4500-a746-3fab7fbfde88\") " pod="openstack/placement-7d689bd78d-s9rtj" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.844344 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6615f806-2d35-4500-a746-3fab7fbfde88-internal-tls-certs\") pod \"placement-7d689bd78d-s9rtj\" (UID: \"6615f806-2d35-4500-a746-3fab7fbfde88\") " pod="openstack/placement-7d689bd78d-s9rtj" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.844400 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0857cf38-58dd-416b-92ca-6b98800ff512-combined-ca-bundle\") pod \"keystone-58576c9895-sk9tb\" (UID: \"0857cf38-58dd-416b-92ca-6b98800ff512\") " pod="openstack/keystone-58576c9895-sk9tb" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.844422 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0857cf38-58dd-416b-92ca-6b98800ff512-scripts\") pod \"keystone-58576c9895-sk9tb\" (UID: \"0857cf38-58dd-416b-92ca-6b98800ff512\") " pod="openstack/keystone-58576c9895-sk9tb" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.844478 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfrgb\" (UniqueName: \"kubernetes.io/projected/0857cf38-58dd-416b-92ca-6b98800ff512-kube-api-access-nfrgb\") pod \"keystone-58576c9895-sk9tb\" (UID: \"0857cf38-58dd-416b-92ca-6b98800ff512\") " pod="openstack/keystone-58576c9895-sk9tb" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.844536 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0857cf38-58dd-416b-92ca-6b98800ff512-fernet-keys\") pod \"keystone-58576c9895-sk9tb\" (UID: \"0857cf38-58dd-416b-92ca-6b98800ff512\") " pod="openstack/keystone-58576c9895-sk9tb" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.844567 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6615f806-2d35-4500-a746-3fab7fbfde88-scripts\") pod \"placement-7d689bd78d-s9rtj\" (UID: \"6615f806-2d35-4500-a746-3fab7fbfde88\") " pod="openstack/placement-7d689bd78d-s9rtj" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.844586 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6615f806-2d35-4500-a746-3fab7fbfde88-public-tls-certs\") pod \"placement-7d689bd78d-s9rtj\" (UID: \"6615f806-2d35-4500-a746-3fab7fbfde88\") " pod="openstack/placement-7d689bd78d-s9rtj" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.844607 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvn7b\" (UniqueName: \"kubernetes.io/projected/6615f806-2d35-4500-a746-3fab7fbfde88-kube-api-access-wvn7b\") pod \"placement-7d689bd78d-s9rtj\" (UID: \"6615f806-2d35-4500-a746-3fab7fbfde88\") " pod="openstack/placement-7d689bd78d-s9rtj" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.844658 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0857cf38-58dd-416b-92ca-6b98800ff512-credential-keys\") pod \"keystone-58576c9895-sk9tb\" (UID: \"0857cf38-58dd-416b-92ca-6b98800ff512\") " pod="openstack/keystone-58576c9895-sk9tb" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.844714 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0857cf38-58dd-416b-92ca-6b98800ff512-config-data\") pod \"keystone-58576c9895-sk9tb\" (UID: \"0857cf38-58dd-416b-92ca-6b98800ff512\") " pod="openstack/keystone-58576c9895-sk9tb" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.844747 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6615f806-2d35-4500-a746-3fab7fbfde88-config-data\") pod \"placement-7d689bd78d-s9rtj\" (UID: \"6615f806-2d35-4500-a746-3fab7fbfde88\") " pod="openstack/placement-7d689bd78d-s9rtj" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.844778 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0857cf38-58dd-416b-92ca-6b98800ff512-public-tls-certs\") pod \"keystone-58576c9895-sk9tb\" (UID: \"0857cf38-58dd-416b-92ca-6b98800ff512\") " pod="openstack/keystone-58576c9895-sk9tb" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.844820 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0857cf38-58dd-416b-92ca-6b98800ff512-internal-tls-certs\") pod \"keystone-58576c9895-sk9tb\" (UID: \"0857cf38-58dd-416b-92ca-6b98800ff512\") " pod="openstack/keystone-58576c9895-sk9tb" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.844875 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6615f806-2d35-4500-a746-3fab7fbfde88-combined-ca-bundle\") pod \"placement-7d689bd78d-s9rtj\" (UID: \"6615f806-2d35-4500-a746-3fab7fbfde88\") " pod="openstack/placement-7d689bd78d-s9rtj" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.861044 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0857cf38-58dd-416b-92ca-6b98800ff512-public-tls-certs\") pod \"keystone-58576c9895-sk9tb\" (UID: \"0857cf38-58dd-416b-92ca-6b98800ff512\") " pod="openstack/keystone-58576c9895-sk9tb" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.863793 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0857cf38-58dd-416b-92ca-6b98800ff512-config-data\") pod \"keystone-58576c9895-sk9tb\" (UID: \"0857cf38-58dd-416b-92ca-6b98800ff512\") " pod="openstack/keystone-58576c9895-sk9tb" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.864458 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0857cf38-58dd-416b-92ca-6b98800ff512-internal-tls-certs\") pod \"keystone-58576c9895-sk9tb\" (UID: \"0857cf38-58dd-416b-92ca-6b98800ff512\") " pod="openstack/keystone-58576c9895-sk9tb" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.869185 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0857cf38-58dd-416b-92ca-6b98800ff512-combined-ca-bundle\") pod \"keystone-58576c9895-sk9tb\" (UID: \"0857cf38-58dd-416b-92ca-6b98800ff512\") " pod="openstack/keystone-58576c9895-sk9tb" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.875689 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0857cf38-58dd-416b-92ca-6b98800ff512-fernet-keys\") pod \"keystone-58576c9895-sk9tb\" (UID: \"0857cf38-58dd-416b-92ca-6b98800ff512\") " pod="openstack/keystone-58576c9895-sk9tb" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.876034 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0857cf38-58dd-416b-92ca-6b98800ff512-scripts\") pod \"keystone-58576c9895-sk9tb\" (UID: \"0857cf38-58dd-416b-92ca-6b98800ff512\") " pod="openstack/keystone-58576c9895-sk9tb" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.878326 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfrgb\" (UniqueName: \"kubernetes.io/projected/0857cf38-58dd-416b-92ca-6b98800ff512-kube-api-access-nfrgb\") pod \"keystone-58576c9895-sk9tb\" (UID: \"0857cf38-58dd-416b-92ca-6b98800ff512\") " pod="openstack/keystone-58576c9895-sk9tb" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.879805 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0857cf38-58dd-416b-92ca-6b98800ff512-credential-keys\") pod \"keystone-58576c9895-sk9tb\" (UID: \"0857cf38-58dd-416b-92ca-6b98800ff512\") " pod="openstack/keystone-58576c9895-sk9tb" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.946136 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6615f806-2d35-4500-a746-3fab7fbfde88-combined-ca-bundle\") pod \"placement-7d689bd78d-s9rtj\" (UID: \"6615f806-2d35-4500-a746-3fab7fbfde88\") " pod="openstack/placement-7d689bd78d-s9rtj" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.946247 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6615f806-2d35-4500-a746-3fab7fbfde88-logs\") pod \"placement-7d689bd78d-s9rtj\" (UID: \"6615f806-2d35-4500-a746-3fab7fbfde88\") " pod="openstack/placement-7d689bd78d-s9rtj" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.946490 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6615f806-2d35-4500-a746-3fab7fbfde88-internal-tls-certs\") pod \"placement-7d689bd78d-s9rtj\" (UID: \"6615f806-2d35-4500-a746-3fab7fbfde88\") " pod="openstack/placement-7d689bd78d-s9rtj" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.946569 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6615f806-2d35-4500-a746-3fab7fbfde88-scripts\") pod \"placement-7d689bd78d-s9rtj\" (UID: \"6615f806-2d35-4500-a746-3fab7fbfde88\") " pod="openstack/placement-7d689bd78d-s9rtj" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.946591 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6615f806-2d35-4500-a746-3fab7fbfde88-public-tls-certs\") pod \"placement-7d689bd78d-s9rtj\" (UID: \"6615f806-2d35-4500-a746-3fab7fbfde88\") " pod="openstack/placement-7d689bd78d-s9rtj" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.946613 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvn7b\" (UniqueName: \"kubernetes.io/projected/6615f806-2d35-4500-a746-3fab7fbfde88-kube-api-access-wvn7b\") pod \"placement-7d689bd78d-s9rtj\" (UID: \"6615f806-2d35-4500-a746-3fab7fbfde88\") " pod="openstack/placement-7d689bd78d-s9rtj" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.946675 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6615f806-2d35-4500-a746-3fab7fbfde88-config-data\") pod \"placement-7d689bd78d-s9rtj\" (UID: \"6615f806-2d35-4500-a746-3fab7fbfde88\") " pod="openstack/placement-7d689bd78d-s9rtj" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.949370 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6615f806-2d35-4500-a746-3fab7fbfde88-logs\") pod \"placement-7d689bd78d-s9rtj\" (UID: \"6615f806-2d35-4500-a746-3fab7fbfde88\") " pod="openstack/placement-7d689bd78d-s9rtj" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.951431 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6615f806-2d35-4500-a746-3fab7fbfde88-config-data\") pod \"placement-7d689bd78d-s9rtj\" (UID: \"6615f806-2d35-4500-a746-3fab7fbfde88\") " pod="openstack/placement-7d689bd78d-s9rtj" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.953797 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6615f806-2d35-4500-a746-3fab7fbfde88-internal-tls-certs\") pod \"placement-7d689bd78d-s9rtj\" (UID: \"6615f806-2d35-4500-a746-3fab7fbfde88\") " pod="openstack/placement-7d689bd78d-s9rtj" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.954881 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6615f806-2d35-4500-a746-3fab7fbfde88-combined-ca-bundle\") pod \"placement-7d689bd78d-s9rtj\" (UID: \"6615f806-2d35-4500-a746-3fab7fbfde88\") " pod="openstack/placement-7d689bd78d-s9rtj" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.955676 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6615f806-2d35-4500-a746-3fab7fbfde88-public-tls-certs\") pod \"placement-7d689bd78d-s9rtj\" (UID: \"6615f806-2d35-4500-a746-3fab7fbfde88\") " pod="openstack/placement-7d689bd78d-s9rtj" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.975691 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6615f806-2d35-4500-a746-3fab7fbfde88-scripts\") pod \"placement-7d689bd78d-s9rtj\" (UID: \"6615f806-2d35-4500-a746-3fab7fbfde88\") " pod="openstack/placement-7d689bd78d-s9rtj" Dec 01 08:59:55 crc kubenswrapper[4873]: I1201 08:59:55.982301 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvn7b\" (UniqueName: \"kubernetes.io/projected/6615f806-2d35-4500-a746-3fab7fbfde88-kube-api-access-wvn7b\") pod \"placement-7d689bd78d-s9rtj\" (UID: \"6615f806-2d35-4500-a746-3fab7fbfde88\") " pod="openstack/placement-7d689bd78d-s9rtj" Dec 01 08:59:56 crc kubenswrapper[4873]: I1201 08:59:56.129331 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-58576c9895-sk9tb" Dec 01 08:59:56 crc kubenswrapper[4873]: I1201 08:59:56.166096 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7d689bd78d-s9rtj" Dec 01 08:59:56 crc kubenswrapper[4873]: I1201 08:59:56.685691 4873 generic.go:334] "Generic (PLEG): container finished" podID="21098264-17f0-46eb-8531-7df74404e5c7" containerID="40337c1bdcc463c9337b54bce609af965f1439ed1e6d6a8a3ae9b552e17704d2" exitCode=0 Dec 01 08:59:56 crc kubenswrapper[4873]: I1201 08:59:56.720389 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-f99dddf57-sr27n" podStartSLOduration=3.7203677109999997 podStartE2EDuration="3.720367711s" podCreationTimestamp="2025-12-01 08:59:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:59:56.719818838 +0000 UTC m=+1172.621927377" watchObservedRunningTime="2025-12-01 08:59:56.720367711 +0000 UTC m=+1172.622476250" Dec 01 08:59:56 crc kubenswrapper[4873]: I1201 08:59:56.720777 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f99dddf57-sr27n" event={"ID":"7b48bbdb-ef25-4a25-a79f-d01eca3d63df","Type":"ContainerStarted","Data":"dc939d0d5ade61cfe01229f129c0b841d48b90fac3abf4198f9de12ea6c27edb"} Dec 01 08:59:56 crc kubenswrapper[4873]: I1201 08:59:56.720836 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-f99dddf57-sr27n" Dec 01 08:59:56 crc kubenswrapper[4873]: I1201 08:59:56.720848 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f99dddf57-sr27n" event={"ID":"7b48bbdb-ef25-4a25-a79f-d01eca3d63df","Type":"ContainerStarted","Data":"96d818107c2ac498bfe1852a6edb2d73dff36d3e6df23acc03cd0f52d3dbaf3f"} Dec 01 08:59:56 crc kubenswrapper[4873]: I1201 08:59:56.720860 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-56d597fbb8-m5cgw" Dec 01 08:59:56 crc kubenswrapper[4873]: I1201 08:59:56.720871 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f99dddf57-sr27n" event={"ID":"7b48bbdb-ef25-4a25-a79f-d01eca3d63df","Type":"ContainerStarted","Data":"dab10eb2e00f73d7880b4a70b4e0a54f5b7e47e099cbe96b25735a91f45b9f08"} Dec 01 08:59:56 crc kubenswrapper[4873]: I1201 08:59:56.720882 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b946d459c-csc2t" event={"ID":"21098264-17f0-46eb-8531-7df74404e5c7","Type":"ContainerDied","Data":"40337c1bdcc463c9337b54bce609af965f1439ed1e6d6a8a3ae9b552e17704d2"} Dec 01 08:59:56 crc kubenswrapper[4873]: I1201 08:59:56.720897 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b946d459c-csc2t" event={"ID":"21098264-17f0-46eb-8531-7df74404e5c7","Type":"ContainerStarted","Data":"0f59effb39fc254d1509c8e1d2f0dbbda12836bda1f8d041eaf4e461110409f9"} Dec 01 08:59:56 crc kubenswrapper[4873]: I1201 08:59:56.720907 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-56d597fbb8-m5cgw" event={"ID":"1a53851f-14c9-4577-bb4e-c3a7b6b8c141","Type":"ContainerStarted","Data":"86c8b1a2e01ed1ec8a465652759bb8d665cb8b35cb8ff4cd32090ac977656e38"} Dec 01 08:59:56 crc kubenswrapper[4873]: I1201 08:59:56.720918 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-56d597fbb8-m5cgw" event={"ID":"1a53851f-14c9-4577-bb4e-c3a7b6b8c141","Type":"ContainerStarted","Data":"775289bce97005f19b4ef7e8bc7d99f9fad1ebab676f25f563b8efaf92232c61"} Dec 01 08:59:56 crc kubenswrapper[4873]: I1201 08:59:56.720927 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-56d597fbb8-m5cgw" event={"ID":"1a53851f-14c9-4577-bb4e-c3a7b6b8c141","Type":"ContainerStarted","Data":"dac62a9725f74be0188bcea687e04bdd8d000362222db48caba4d305358af664"} Dec 01 08:59:56 crc kubenswrapper[4873]: I1201 08:59:56.767643 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-56d597fbb8-m5cgw" podStartSLOduration=6.767445922 podStartE2EDuration="6.767445922s" podCreationTimestamp="2025-12-01 08:59:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:59:56.761388704 +0000 UTC m=+1172.663497243" watchObservedRunningTime="2025-12-01 08:59:56.767445922 +0000 UTC m=+1172.669554461" Dec 01 08:59:57 crc kubenswrapper[4873]: I1201 08:59:57.080202 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7d689bd78d-s9rtj"] Dec 01 08:59:57 crc kubenswrapper[4873]: I1201 08:59:57.142883 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-58576c9895-sk9tb"] Dec 01 08:59:57 crc kubenswrapper[4873]: I1201 08:59:57.707729 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7d689bd78d-s9rtj" event={"ID":"6615f806-2d35-4500-a746-3fab7fbfde88","Type":"ContainerStarted","Data":"d0736a03ff801bde49ea30501028423e19f255bf8a177b0ab6fafcfdf70a94fc"} Dec 01 08:59:57 crc kubenswrapper[4873]: I1201 08:59:57.709881 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7d689bd78d-s9rtj" event={"ID":"6615f806-2d35-4500-a746-3fab7fbfde88","Type":"ContainerStarted","Data":"b8b913d6e617e694c779ec76053a362e336a3ecd7e2807746a6d578bc60c2214"} Dec 01 08:59:57 crc kubenswrapper[4873]: I1201 08:59:57.709994 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-58576c9895-sk9tb" event={"ID":"0857cf38-58dd-416b-92ca-6b98800ff512","Type":"ContainerStarted","Data":"80f59ab6527208fd03b93a4db9a666b878b4c05c898875cac5bcfd8b3e7e0f70"} Dec 01 08:59:57 crc kubenswrapper[4873]: I1201 08:59:57.710120 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-58576c9895-sk9tb" event={"ID":"0857cf38-58dd-416b-92ca-6b98800ff512","Type":"ContainerStarted","Data":"5a401d33dde1213488780b31dd95bac28470603c68ab3610fb585d7eb07794f3"} Dec 01 08:59:57 crc kubenswrapper[4873]: I1201 08:59:57.711099 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b946d459c-csc2t" event={"ID":"21098264-17f0-46eb-8531-7df74404e5c7","Type":"ContainerStarted","Data":"ec850baf9aac29ca65fec4acd31d5cbcdff67cf2b06600c086dd67f67b4840f7"} Dec 01 08:59:57 crc kubenswrapper[4873]: I1201 08:59:57.744512 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7b946d459c-csc2t" podStartSLOduration=7.744492057 podStartE2EDuration="7.744492057s" podCreationTimestamp="2025-12-01 08:59:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:59:57.733904678 +0000 UTC m=+1173.636013217" watchObservedRunningTime="2025-12-01 08:59:57.744492057 +0000 UTC m=+1173.646600596" Dec 01 08:59:58 crc kubenswrapper[4873]: I1201 08:59:58.718472 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7b946d459c-csc2t" Dec 01 08:59:59 crc kubenswrapper[4873]: I1201 08:59:59.739720 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7d689bd78d-s9rtj" event={"ID":"6615f806-2d35-4500-a746-3fab7fbfde88","Type":"ContainerStarted","Data":"40aa5b94a539e5c632cc848792529134b9b5734d26f6312cd2ce0a02fe8fa14c"} Dec 01 08:59:59 crc kubenswrapper[4873]: I1201 08:59:59.740464 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7d689bd78d-s9rtj" Dec 01 08:59:59 crc kubenswrapper[4873]: I1201 08:59:59.740493 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7d689bd78d-s9rtj" Dec 01 08:59:59 crc kubenswrapper[4873]: I1201 08:59:59.746988 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-58576c9895-sk9tb" Dec 01 08:59:59 crc kubenswrapper[4873]: I1201 08:59:59.779980 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-7d689bd78d-s9rtj" podStartSLOduration=4.779955095 podStartE2EDuration="4.779955095s" podCreationTimestamp="2025-12-01 08:59:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:59:59.770063493 +0000 UTC m=+1175.672172052" watchObservedRunningTime="2025-12-01 08:59:59.779955095 +0000 UTC m=+1175.682063634" Dec 01 08:59:59 crc kubenswrapper[4873]: I1201 08:59:59.800469 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-58576c9895-sk9tb" podStartSLOduration=4.800442346 podStartE2EDuration="4.800442346s" podCreationTimestamp="2025-12-01 08:59:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:59:59.799510564 +0000 UTC m=+1175.701619113" watchObservedRunningTime="2025-12-01 08:59:59.800442346 +0000 UTC m=+1175.702550885" Dec 01 09:00:00 crc kubenswrapper[4873]: I1201 09:00:00.159246 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409660-hsrhk"] Dec 01 09:00:00 crc kubenswrapper[4873]: I1201 09:00:00.173508 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409660-hsrhk"] Dec 01 09:00:00 crc kubenswrapper[4873]: I1201 09:00:00.173730 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-hsrhk" Dec 01 09:00:00 crc kubenswrapper[4873]: I1201 09:00:00.178993 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 09:00:00 crc kubenswrapper[4873]: I1201 09:00:00.179646 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 09:00:00 crc kubenswrapper[4873]: I1201 09:00:00.305665 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/467ecd19-95ef-4663-b8bd-0a5966bceb01-secret-volume\") pod \"collect-profiles-29409660-hsrhk\" (UID: \"467ecd19-95ef-4663-b8bd-0a5966bceb01\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-hsrhk" Dec 01 09:00:00 crc kubenswrapper[4873]: I1201 09:00:00.305778 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/467ecd19-95ef-4663-b8bd-0a5966bceb01-config-volume\") pod \"collect-profiles-29409660-hsrhk\" (UID: \"467ecd19-95ef-4663-b8bd-0a5966bceb01\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-hsrhk" Dec 01 09:00:00 crc kubenswrapper[4873]: I1201 09:00:00.305804 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8l74g\" (UniqueName: \"kubernetes.io/projected/467ecd19-95ef-4663-b8bd-0a5966bceb01-kube-api-access-8l74g\") pod \"collect-profiles-29409660-hsrhk\" (UID: \"467ecd19-95ef-4663-b8bd-0a5966bceb01\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-hsrhk" Dec 01 09:00:00 crc kubenswrapper[4873]: I1201 09:00:00.408129 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/467ecd19-95ef-4663-b8bd-0a5966bceb01-secret-volume\") pod \"collect-profiles-29409660-hsrhk\" (UID: \"467ecd19-95ef-4663-b8bd-0a5966bceb01\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-hsrhk" Dec 01 09:00:00 crc kubenswrapper[4873]: I1201 09:00:00.408301 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/467ecd19-95ef-4663-b8bd-0a5966bceb01-config-volume\") pod \"collect-profiles-29409660-hsrhk\" (UID: \"467ecd19-95ef-4663-b8bd-0a5966bceb01\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-hsrhk" Dec 01 09:00:00 crc kubenswrapper[4873]: I1201 09:00:00.408328 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8l74g\" (UniqueName: \"kubernetes.io/projected/467ecd19-95ef-4663-b8bd-0a5966bceb01-kube-api-access-8l74g\") pod \"collect-profiles-29409660-hsrhk\" (UID: \"467ecd19-95ef-4663-b8bd-0a5966bceb01\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-hsrhk" Dec 01 09:00:00 crc kubenswrapper[4873]: I1201 09:00:00.411901 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/467ecd19-95ef-4663-b8bd-0a5966bceb01-config-volume\") pod \"collect-profiles-29409660-hsrhk\" (UID: \"467ecd19-95ef-4663-b8bd-0a5966bceb01\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-hsrhk" Dec 01 09:00:00 crc kubenswrapper[4873]: I1201 09:00:00.416684 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/467ecd19-95ef-4663-b8bd-0a5966bceb01-secret-volume\") pod \"collect-profiles-29409660-hsrhk\" (UID: \"467ecd19-95ef-4663-b8bd-0a5966bceb01\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-hsrhk" Dec 01 09:00:00 crc kubenswrapper[4873]: I1201 09:00:00.433518 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8l74g\" (UniqueName: \"kubernetes.io/projected/467ecd19-95ef-4663-b8bd-0a5966bceb01-kube-api-access-8l74g\") pod \"collect-profiles-29409660-hsrhk\" (UID: \"467ecd19-95ef-4663-b8bd-0a5966bceb01\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-hsrhk" Dec 01 09:00:00 crc kubenswrapper[4873]: I1201 09:00:00.517518 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-hsrhk" Dec 01 09:00:00 crc kubenswrapper[4873]: I1201 09:00:00.787986 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-bmhcv" event={"ID":"ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce","Type":"ContainerStarted","Data":"00cb6d0f184dc4c23b88389120c79e311420e96e7e478921739301747dbe7aa5"} Dec 01 09:00:00 crc kubenswrapper[4873]: I1201 09:00:00.806929 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-9lvtd" event={"ID":"c8c5adca-9738-46b1-9376-01cfb425d839","Type":"ContainerStarted","Data":"930b7227d8f1722404cdef589bc743c34ec93d72d99ae635ecdc38fc41b85884"} Dec 01 09:00:00 crc kubenswrapper[4873]: I1201 09:00:00.831103 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-bmhcv" podStartSLOduration=3.261485901 podStartE2EDuration="38.831068781s" podCreationTimestamp="2025-12-01 08:59:22 +0000 UTC" firstStartedPulling="2025-12-01 08:59:23.96807525 +0000 UTC m=+1139.870183799" lastFinishedPulling="2025-12-01 08:59:59.53765814 +0000 UTC m=+1175.439766679" observedRunningTime="2025-12-01 09:00:00.81710161 +0000 UTC m=+1176.719210159" watchObservedRunningTime="2025-12-01 09:00:00.831068781 +0000 UTC m=+1176.733177330" Dec 01 09:00:00 crc kubenswrapper[4873]: I1201 09:00:00.866964 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-9lvtd" podStartSLOduration=3.76060467 podStartE2EDuration="38.866930668s" podCreationTimestamp="2025-12-01 08:59:22 +0000 UTC" firstStartedPulling="2025-12-01 08:59:24.261786356 +0000 UTC m=+1140.163894895" lastFinishedPulling="2025-12-01 08:59:59.368112354 +0000 UTC m=+1175.270220893" observedRunningTime="2025-12-01 09:00:00.839898817 +0000 UTC m=+1176.742007356" watchObservedRunningTime="2025-12-01 09:00:00.866930668 +0000 UTC m=+1176.769039207" Dec 01 09:00:01 crc kubenswrapper[4873]: I1201 09:00:01.078631 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409660-hsrhk"] Dec 01 09:00:01 crc kubenswrapper[4873]: I1201 09:00:01.820538 4873 generic.go:334] "Generic (PLEG): container finished" podID="467ecd19-95ef-4663-b8bd-0a5966bceb01" containerID="e7dafb6465a448337fa1b3df8996799cf858e394f863a00d25c0e5d27063e846" exitCode=0 Dec 01 09:00:01 crc kubenswrapper[4873]: I1201 09:00:01.820596 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-hsrhk" event={"ID":"467ecd19-95ef-4663-b8bd-0a5966bceb01","Type":"ContainerDied","Data":"e7dafb6465a448337fa1b3df8996799cf858e394f863a00d25c0e5d27063e846"} Dec 01 09:00:01 crc kubenswrapper[4873]: I1201 09:00:01.821051 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-hsrhk" event={"ID":"467ecd19-95ef-4663-b8bd-0a5966bceb01","Type":"ContainerStarted","Data":"bc389b94ff20eb255e47e660d42681eac1ddf51a3aba75c4f90ded8f4e69ff6c"} Dec 01 09:00:03 crc kubenswrapper[4873]: I1201 09:00:03.850432 4873 generic.go:334] "Generic (PLEG): container finished" podID="c8c5adca-9738-46b1-9376-01cfb425d839" containerID="930b7227d8f1722404cdef589bc743c34ec93d72d99ae635ecdc38fc41b85884" exitCode=0 Dec 01 09:00:03 crc kubenswrapper[4873]: I1201 09:00:03.850993 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-9lvtd" event={"ID":"c8c5adca-9738-46b1-9376-01cfb425d839","Type":"ContainerDied","Data":"930b7227d8f1722404cdef589bc743c34ec93d72d99ae635ecdc38fc41b85884"} Dec 01 09:00:05 crc kubenswrapper[4873]: I1201 09:00:05.853739 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7b946d459c-csc2t" Dec 01 09:00:05 crc kubenswrapper[4873]: I1201 09:00:05.881919 4873 generic.go:334] "Generic (PLEG): container finished" podID="ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce" containerID="00cb6d0f184dc4c23b88389120c79e311420e96e7e478921739301747dbe7aa5" exitCode=0 Dec 01 09:00:05 crc kubenswrapper[4873]: I1201 09:00:05.882636 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-bmhcv" event={"ID":"ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce","Type":"ContainerDied","Data":"00cb6d0f184dc4c23b88389120c79e311420e96e7e478921739301747dbe7aa5"} Dec 01 09:00:05 crc kubenswrapper[4873]: I1201 09:00:05.914757 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7987f74bbc-6rfmd"] Dec 01 09:00:05 crc kubenswrapper[4873]: I1201 09:00:05.915171 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7987f74bbc-6rfmd" podUID="4f027d3a-e35f-418b-bc8b-508a637f6698" containerName="dnsmasq-dns" containerID="cri-o://63a7ee8c78eab560e2b238cb614c249216883b57b716a3838d82ba9f212e2127" gracePeriod=10 Dec 01 09:00:06 crc kubenswrapper[4873]: I1201 09:00:06.897300 4873 generic.go:334] "Generic (PLEG): container finished" podID="4f027d3a-e35f-418b-bc8b-508a637f6698" containerID="63a7ee8c78eab560e2b238cb614c249216883b57b716a3838d82ba9f212e2127" exitCode=0 Dec 01 09:00:06 crc kubenswrapper[4873]: I1201 09:00:06.897360 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7987f74bbc-6rfmd" event={"ID":"4f027d3a-e35f-418b-bc8b-508a637f6698","Type":"ContainerDied","Data":"63a7ee8c78eab560e2b238cb614c249216883b57b716a3838d82ba9f212e2127"} Dec 01 09:00:07 crc kubenswrapper[4873]: I1201 09:00:07.766467 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-9lvtd" Dec 01 09:00:07 crc kubenswrapper[4873]: I1201 09:00:07.779283 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-bmhcv" Dec 01 09:00:07 crc kubenswrapper[4873]: I1201 09:00:07.779424 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-hsrhk" Dec 01 09:00:07 crc kubenswrapper[4873]: I1201 09:00:07.907426 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce-etc-machine-id\") pod \"ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce\" (UID: \"ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce\") " Dec 01 09:00:07 crc kubenswrapper[4873]: I1201 09:00:07.907506 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce-combined-ca-bundle\") pod \"ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce\" (UID: \"ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce\") " Dec 01 09:00:07 crc kubenswrapper[4873]: I1201 09:00:07.907683 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce-config-data\") pod \"ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce\" (UID: \"ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce\") " Dec 01 09:00:07 crc kubenswrapper[4873]: I1201 09:00:07.907726 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8l74g\" (UniqueName: \"kubernetes.io/projected/467ecd19-95ef-4663-b8bd-0a5966bceb01-kube-api-access-8l74g\") pod \"467ecd19-95ef-4663-b8bd-0a5966bceb01\" (UID: \"467ecd19-95ef-4663-b8bd-0a5966bceb01\") " Dec 01 09:00:07 crc kubenswrapper[4873]: I1201 09:00:07.907750 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce-db-sync-config-data\") pod \"ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce\" (UID: \"ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce\") " Dec 01 09:00:07 crc kubenswrapper[4873]: I1201 09:00:07.907845 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8c5adca-9738-46b1-9376-01cfb425d839-combined-ca-bundle\") pod \"c8c5adca-9738-46b1-9376-01cfb425d839\" (UID: \"c8c5adca-9738-46b1-9376-01cfb425d839\") " Dec 01 09:00:07 crc kubenswrapper[4873]: I1201 09:00:07.907876 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rtl44\" (UniqueName: \"kubernetes.io/projected/ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce-kube-api-access-rtl44\") pod \"ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce\" (UID: \"ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce\") " Dec 01 09:00:07 crc kubenswrapper[4873]: I1201 09:00:07.907903 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h64g6\" (UniqueName: \"kubernetes.io/projected/c8c5adca-9738-46b1-9376-01cfb425d839-kube-api-access-h64g6\") pod \"c8c5adca-9738-46b1-9376-01cfb425d839\" (UID: \"c8c5adca-9738-46b1-9376-01cfb425d839\") " Dec 01 09:00:07 crc kubenswrapper[4873]: I1201 09:00:07.907949 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce-scripts\") pod \"ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce\" (UID: \"ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce\") " Dec 01 09:00:07 crc kubenswrapper[4873]: I1201 09:00:07.907979 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c8c5adca-9738-46b1-9376-01cfb425d839-db-sync-config-data\") pod \"c8c5adca-9738-46b1-9376-01cfb425d839\" (UID: \"c8c5adca-9738-46b1-9376-01cfb425d839\") " Dec 01 09:00:07 crc kubenswrapper[4873]: I1201 09:00:07.908030 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/467ecd19-95ef-4663-b8bd-0a5966bceb01-config-volume\") pod \"467ecd19-95ef-4663-b8bd-0a5966bceb01\" (UID: \"467ecd19-95ef-4663-b8bd-0a5966bceb01\") " Dec 01 09:00:07 crc kubenswrapper[4873]: I1201 09:00:07.908055 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/467ecd19-95ef-4663-b8bd-0a5966bceb01-secret-volume\") pod \"467ecd19-95ef-4663-b8bd-0a5966bceb01\" (UID: \"467ecd19-95ef-4663-b8bd-0a5966bceb01\") " Dec 01 09:00:07 crc kubenswrapper[4873]: I1201 09:00:07.908851 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce" (UID: "ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:00:07 crc kubenswrapper[4873]: I1201 09:00:07.909170 4873 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:07 crc kubenswrapper[4873]: I1201 09:00:07.913928 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8c5adca-9738-46b1-9376-01cfb425d839-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "c8c5adca-9738-46b1-9376-01cfb425d839" (UID: "c8c5adca-9738-46b1-9376-01cfb425d839"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:07 crc kubenswrapper[4873]: I1201 09:00:07.914047 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/467ecd19-95ef-4663-b8bd-0a5966bceb01-config-volume" (OuterVolumeSpecName: "config-volume") pod "467ecd19-95ef-4663-b8bd-0a5966bceb01" (UID: "467ecd19-95ef-4663-b8bd-0a5966bceb01"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:00:07 crc kubenswrapper[4873]: I1201 09:00:07.914862 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce-scripts" (OuterVolumeSpecName: "scripts") pod "ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce" (UID: "ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:07 crc kubenswrapper[4873]: I1201 09:00:07.918747 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/467ecd19-95ef-4663-b8bd-0a5966bceb01-kube-api-access-8l74g" (OuterVolumeSpecName: "kube-api-access-8l74g") pod "467ecd19-95ef-4663-b8bd-0a5966bceb01" (UID: "467ecd19-95ef-4663-b8bd-0a5966bceb01"). InnerVolumeSpecName "kube-api-access-8l74g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:00:07 crc kubenswrapper[4873]: I1201 09:00:07.921036 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8c5adca-9738-46b1-9376-01cfb425d839-kube-api-access-h64g6" (OuterVolumeSpecName: "kube-api-access-h64g6") pod "c8c5adca-9738-46b1-9376-01cfb425d839" (UID: "c8c5adca-9738-46b1-9376-01cfb425d839"). InnerVolumeSpecName "kube-api-access-h64g6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:00:07 crc kubenswrapper[4873]: I1201 09:00:07.924073 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce-kube-api-access-rtl44" (OuterVolumeSpecName: "kube-api-access-rtl44") pod "ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce" (UID: "ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce"). InnerVolumeSpecName "kube-api-access-rtl44". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:00:07 crc kubenswrapper[4873]: I1201 09:00:07.927109 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-hsrhk" event={"ID":"467ecd19-95ef-4663-b8bd-0a5966bceb01","Type":"ContainerDied","Data":"bc389b94ff20eb255e47e660d42681eac1ddf51a3aba75c4f90ded8f4e69ff6c"} Dec 01 09:00:07 crc kubenswrapper[4873]: I1201 09:00:07.927155 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc389b94ff20eb255e47e660d42681eac1ddf51a3aba75c4f90ded8f4e69ff6c" Dec 01 09:00:07 crc kubenswrapper[4873]: I1201 09:00:07.927243 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-hsrhk" Dec 01 09:00:07 crc kubenswrapper[4873]: I1201 09:00:07.929661 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-bmhcv" Dec 01 09:00:07 crc kubenswrapper[4873]: I1201 09:00:07.929668 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-bmhcv" event={"ID":"ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce","Type":"ContainerDied","Data":"17d905efe968146eb4a57a4bd628436f696ff2dd64309260dce5d653dde313b6"} Dec 01 09:00:07 crc kubenswrapper[4873]: I1201 09:00:07.929723 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="17d905efe968146eb4a57a4bd628436f696ff2dd64309260dce5d653dde313b6" Dec 01 09:00:07 crc kubenswrapper[4873]: I1201 09:00:07.943315 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/467ecd19-95ef-4663-b8bd-0a5966bceb01-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "467ecd19-95ef-4663-b8bd-0a5966bceb01" (UID: "467ecd19-95ef-4663-b8bd-0a5966bceb01"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:07 crc kubenswrapper[4873]: I1201 09:00:07.945617 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce" (UID: "ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:07 crc kubenswrapper[4873]: I1201 09:00:07.958780 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-9lvtd" event={"ID":"c8c5adca-9738-46b1-9376-01cfb425d839","Type":"ContainerDied","Data":"cec67fe9cd51f0192a8e60ce1cea570ee2e34eaa68a29440df00512e6246210f"} Dec 01 09:00:07 crc kubenswrapper[4873]: I1201 09:00:07.958829 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cec67fe9cd51f0192a8e60ce1cea570ee2e34eaa68a29440df00512e6246210f" Dec 01 09:00:07 crc kubenswrapper[4873]: I1201 09:00:07.958947 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-9lvtd" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.003179 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce-config-data" (OuterVolumeSpecName: "config-data") pod "ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce" (UID: "ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.010288 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce" (UID: "ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.010957 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce-combined-ca-bundle\") pod \"ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce\" (UID: \"ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce\") " Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.011113 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8c5adca-9738-46b1-9376-01cfb425d839-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c8c5adca-9738-46b1-9376-01cfb425d839" (UID: "c8c5adca-9738-46b1-9376-01cfb425d839"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:08 crc kubenswrapper[4873]: W1201 09:00:08.011141 4873 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce/volumes/kubernetes.io~secret/combined-ca-bundle Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.011197 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce" (UID: "ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.011349 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8c5adca-9738-46b1-9376-01cfb425d839-combined-ca-bundle\") pod \"c8c5adca-9738-46b1-9376-01cfb425d839\" (UID: \"c8c5adca-9738-46b1-9376-01cfb425d839\") " Dec 01 09:00:08 crc kubenswrapper[4873]: W1201 09:00:08.011469 4873 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/c8c5adca-9738-46b1-9376-01cfb425d839/volumes/kubernetes.io~secret/combined-ca-bundle Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.011551 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8c5adca-9738-46b1-9376-01cfb425d839-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c8c5adca-9738-46b1-9376-01cfb425d839" (UID: "c8c5adca-9738-46b1-9376-01cfb425d839"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.012041 4873 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/467ecd19-95ef-4663-b8bd-0a5966bceb01-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.012070 4873 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/467ecd19-95ef-4663-b8bd-0a5966bceb01-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.012083 4873 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.012092 4873 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.012128 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8l74g\" (UniqueName: \"kubernetes.io/projected/467ecd19-95ef-4663-b8bd-0a5966bceb01-kube-api-access-8l74g\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.012143 4873 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.012153 4873 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8c5adca-9738-46b1-9376-01cfb425d839-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.012163 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rtl44\" (UniqueName: \"kubernetes.io/projected/ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce-kube-api-access-rtl44\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.012175 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h64g6\" (UniqueName: \"kubernetes.io/projected/c8c5adca-9738-46b1-9376-01cfb425d839-kube-api-access-h64g6\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.012218 4873 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.012232 4873 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c8c5adca-9738-46b1-9376-01cfb425d839-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.068941 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7987f74bbc-6rfmd" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.214941 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-njv2r\" (UniqueName: \"kubernetes.io/projected/4f027d3a-e35f-418b-bc8b-508a637f6698-kube-api-access-njv2r\") pod \"4f027d3a-e35f-418b-bc8b-508a637f6698\" (UID: \"4f027d3a-e35f-418b-bc8b-508a637f6698\") " Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.215115 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4f027d3a-e35f-418b-bc8b-508a637f6698-ovsdbserver-sb\") pod \"4f027d3a-e35f-418b-bc8b-508a637f6698\" (UID: \"4f027d3a-e35f-418b-bc8b-508a637f6698\") " Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.215234 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f027d3a-e35f-418b-bc8b-508a637f6698-dns-svc\") pod \"4f027d3a-e35f-418b-bc8b-508a637f6698\" (UID: \"4f027d3a-e35f-418b-bc8b-508a637f6698\") " Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.215264 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f027d3a-e35f-418b-bc8b-508a637f6698-ovsdbserver-nb\") pod \"4f027d3a-e35f-418b-bc8b-508a637f6698\" (UID: \"4f027d3a-e35f-418b-bc8b-508a637f6698\") " Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.215290 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f027d3a-e35f-418b-bc8b-508a637f6698-config\") pod \"4f027d3a-e35f-418b-bc8b-508a637f6698\" (UID: \"4f027d3a-e35f-418b-bc8b-508a637f6698\") " Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.239298 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f027d3a-e35f-418b-bc8b-508a637f6698-kube-api-access-njv2r" (OuterVolumeSpecName: "kube-api-access-njv2r") pod "4f027d3a-e35f-418b-bc8b-508a637f6698" (UID: "4f027d3a-e35f-418b-bc8b-508a637f6698"). InnerVolumeSpecName "kube-api-access-njv2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.317424 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-njv2r\" (UniqueName: \"kubernetes.io/projected/4f027d3a-e35f-418b-bc8b-508a637f6698-kube-api-access-njv2r\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.368835 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f027d3a-e35f-418b-bc8b-508a637f6698-config" (OuterVolumeSpecName: "config") pod "4f027d3a-e35f-418b-bc8b-508a637f6698" (UID: "4f027d3a-e35f-418b-bc8b-508a637f6698"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.374209 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 09:00:08 crc kubenswrapper[4873]: E1201 09:00:08.374665 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce" containerName="cinder-db-sync" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.374679 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce" containerName="cinder-db-sync" Dec 01 09:00:08 crc kubenswrapper[4873]: E1201 09:00:08.374690 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8c5adca-9738-46b1-9376-01cfb425d839" containerName="barbican-db-sync" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.374696 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8c5adca-9738-46b1-9376-01cfb425d839" containerName="barbican-db-sync" Dec 01 09:00:08 crc kubenswrapper[4873]: E1201 09:00:08.374710 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f027d3a-e35f-418b-bc8b-508a637f6698" containerName="init" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.374717 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f027d3a-e35f-418b-bc8b-508a637f6698" containerName="init" Dec 01 09:00:08 crc kubenswrapper[4873]: E1201 09:00:08.374733 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f027d3a-e35f-418b-bc8b-508a637f6698" containerName="dnsmasq-dns" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.374741 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f027d3a-e35f-418b-bc8b-508a637f6698" containerName="dnsmasq-dns" Dec 01 09:00:08 crc kubenswrapper[4873]: E1201 09:00:08.374761 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="467ecd19-95ef-4663-b8bd-0a5966bceb01" containerName="collect-profiles" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.374768 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="467ecd19-95ef-4663-b8bd-0a5966bceb01" containerName="collect-profiles" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.374964 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8c5adca-9738-46b1-9376-01cfb425d839" containerName="barbican-db-sync" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.374988 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="467ecd19-95ef-4663-b8bd-0a5966bceb01" containerName="collect-profiles" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.375000 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f027d3a-e35f-418b-bc8b-508a637f6698" containerName="dnsmasq-dns" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.375038 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce" containerName="cinder-db-sync" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.376001 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.390204 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.390244 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f027d3a-e35f-418b-bc8b-508a637f6698-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4f027d3a-e35f-418b-bc8b-508a637f6698" (UID: "4f027d3a-e35f-418b-bc8b-508a637f6698"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.390404 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.390455 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-snh6v" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.390624 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.423970 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f027d3a-e35f-418b-bc8b-508a637f6698-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4f027d3a-e35f-418b-bc8b-508a637f6698" (UID: "4f027d3a-e35f-418b-bc8b-508a637f6698"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.425432 4873 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4f027d3a-e35f-418b-bc8b-508a637f6698-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.425453 4873 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f027d3a-e35f-418b-bc8b-508a637f6698-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.425463 4873 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f027d3a-e35f-418b-bc8b-508a637f6698-config\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.426446 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.456856 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f027d3a-e35f-418b-bc8b-508a637f6698-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4f027d3a-e35f-418b-bc8b-508a637f6698" (UID: "4f027d3a-e35f-418b-bc8b-508a637f6698"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.529400 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/11731b74-fe99-40c4-9445-f86102242d05-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"11731b74-fe99-40c4-9445-f86102242d05\") " pod="openstack/cinder-scheduler-0" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.529455 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11731b74-fe99-40c4-9445-f86102242d05-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"11731b74-fe99-40c4-9445-f86102242d05\") " pod="openstack/cinder-scheduler-0" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.529530 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n84tf\" (UniqueName: \"kubernetes.io/projected/11731b74-fe99-40c4-9445-f86102242d05-kube-api-access-n84tf\") pod \"cinder-scheduler-0\" (UID: \"11731b74-fe99-40c4-9445-f86102242d05\") " pod="openstack/cinder-scheduler-0" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.529556 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/11731b74-fe99-40c4-9445-f86102242d05-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"11731b74-fe99-40c4-9445-f86102242d05\") " pod="openstack/cinder-scheduler-0" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.529586 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11731b74-fe99-40c4-9445-f86102242d05-config-data\") pod \"cinder-scheduler-0\" (UID: \"11731b74-fe99-40c4-9445-f86102242d05\") " pod="openstack/cinder-scheduler-0" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.529611 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11731b74-fe99-40c4-9445-f86102242d05-scripts\") pod \"cinder-scheduler-0\" (UID: \"11731b74-fe99-40c4-9445-f86102242d05\") " pod="openstack/cinder-scheduler-0" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.529696 4873 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f027d3a-e35f-418b-bc8b-508a637f6698-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.639422 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n84tf\" (UniqueName: \"kubernetes.io/projected/11731b74-fe99-40c4-9445-f86102242d05-kube-api-access-n84tf\") pod \"cinder-scheduler-0\" (UID: \"11731b74-fe99-40c4-9445-f86102242d05\") " pod="openstack/cinder-scheduler-0" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.639492 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/11731b74-fe99-40c4-9445-f86102242d05-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"11731b74-fe99-40c4-9445-f86102242d05\") " pod="openstack/cinder-scheduler-0" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.639529 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11731b74-fe99-40c4-9445-f86102242d05-config-data\") pod \"cinder-scheduler-0\" (UID: \"11731b74-fe99-40c4-9445-f86102242d05\") " pod="openstack/cinder-scheduler-0" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.639560 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11731b74-fe99-40c4-9445-f86102242d05-scripts\") pod \"cinder-scheduler-0\" (UID: \"11731b74-fe99-40c4-9445-f86102242d05\") " pod="openstack/cinder-scheduler-0" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.639633 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/11731b74-fe99-40c4-9445-f86102242d05-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"11731b74-fe99-40c4-9445-f86102242d05\") " pod="openstack/cinder-scheduler-0" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.642513 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/11731b74-fe99-40c4-9445-f86102242d05-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"11731b74-fe99-40c4-9445-f86102242d05\") " pod="openstack/cinder-scheduler-0" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.640880 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11731b74-fe99-40c4-9445-f86102242d05-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"11731b74-fe99-40c4-9445-f86102242d05\") " pod="openstack/cinder-scheduler-0" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.668130 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/11731b74-fe99-40c4-9445-f86102242d05-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"11731b74-fe99-40c4-9445-f86102242d05\") " pod="openstack/cinder-scheduler-0" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.675114 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11731b74-fe99-40c4-9445-f86102242d05-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"11731b74-fe99-40c4-9445-f86102242d05\") " pod="openstack/cinder-scheduler-0" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.675564 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11731b74-fe99-40c4-9445-f86102242d05-scripts\") pod \"cinder-scheduler-0\" (UID: \"11731b74-fe99-40c4-9445-f86102242d05\") " pod="openstack/cinder-scheduler-0" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.675748 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11731b74-fe99-40c4-9445-f86102242d05-config-data\") pod \"cinder-scheduler-0\" (UID: \"11731b74-fe99-40c4-9445-f86102242d05\") " pod="openstack/cinder-scheduler-0" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.707044 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n84tf\" (UniqueName: \"kubernetes.io/projected/11731b74-fe99-40c4-9445-f86102242d05-kube-api-access-n84tf\") pod \"cinder-scheduler-0\" (UID: \"11731b74-fe99-40c4-9445-f86102242d05\") " pod="openstack/cinder-scheduler-0" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.908289 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f64d5748f-cjt2l"] Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.921429 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f64d5748f-cjt2l"] Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.921467 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.921874 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f64d5748f-cjt2l" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.924170 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.924269 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.928422 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.983109 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e53689a0-5d92-4e62-bb63-3a56af762a59","Type":"ContainerStarted","Data":"068b339c05967459adb9767502db98e3ed7f8097f9d5a942acefdc2ee1b56026"} Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.983384 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e53689a0-5d92-4e62-bb63-3a56af762a59" containerName="ceilometer-central-agent" containerID="cri-o://31940bfd155d788c99f80b4389ebba5de7bb5078293bf263879bda05942db53b" gracePeriod=30 Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.983493 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.983545 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e53689a0-5d92-4e62-bb63-3a56af762a59" containerName="proxy-httpd" containerID="cri-o://068b339c05967459adb9767502db98e3ed7f8097f9d5a942acefdc2ee1b56026" gracePeriod=30 Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.983608 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e53689a0-5d92-4e62-bb63-3a56af762a59" containerName="sg-core" containerID="cri-o://b33b0f231101926f73eefa68214880e4c14df8323cb5ce1c1750a3ab4b4e019d" gracePeriod=30 Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.983650 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e53689a0-5d92-4e62-bb63-3a56af762a59" containerName="ceilometer-notification-agent" containerID="cri-o://0b43d48616acc5de469072536f3377110552ad3499b67b694a963f6c3a326cbc" gracePeriod=30 Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.997004 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7987f74bbc-6rfmd" event={"ID":"4f027d3a-e35f-418b-bc8b-508a637f6698","Type":"ContainerDied","Data":"c752c8f201dda9c6baeeb62e7ac13a45eaf39e4b2157221faff6eb0f7b243d74"} Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.997124 4873 scope.go:117] "RemoveContainer" containerID="63a7ee8c78eab560e2b238cb614c249216883b57b716a3838d82ba9f212e2127" Dec 01 09:00:08 crc kubenswrapper[4873]: I1201 09:00:08.997277 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7987f74bbc-6rfmd" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.047868 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.812087296 podStartE2EDuration="47.047843258s" podCreationTimestamp="2025-12-01 08:59:22 +0000 UTC" firstStartedPulling="2025-12-01 08:59:24.517498335 +0000 UTC m=+1140.419606874" lastFinishedPulling="2025-12-01 09:00:07.753254297 +0000 UTC m=+1183.655362836" observedRunningTime="2025-12-01 09:00:09.021951954 +0000 UTC m=+1184.924060503" watchObservedRunningTime="2025-12-01 09:00:09.047843258 +0000 UTC m=+1184.949951787" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.058213 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b9a7d7f4-a682-4e90-99e4-b3b2afe2c718-etc-machine-id\") pod \"cinder-api-0\" (UID: \"b9a7d7f4-a682-4e90-99e4-b3b2afe2c718\") " pod="openstack/cinder-api-0" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.058263 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swj9t\" (UniqueName: \"kubernetes.io/projected/b9a7d7f4-a682-4e90-99e4-b3b2afe2c718-kube-api-access-swj9t\") pod \"cinder-api-0\" (UID: \"b9a7d7f4-a682-4e90-99e4-b3b2afe2c718\") " pod="openstack/cinder-api-0" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.058309 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b9a7d7f4-a682-4e90-99e4-b3b2afe2c718-config-data-custom\") pod \"cinder-api-0\" (UID: \"b9a7d7f4-a682-4e90-99e4-b3b2afe2c718\") " pod="openstack/cinder-api-0" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.058329 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d504ddc2-fb6b-4307-9380-3c4fd24ca8db-config\") pod \"dnsmasq-dns-f64d5748f-cjt2l\" (UID: \"d504ddc2-fb6b-4307-9380-3c4fd24ca8db\") " pod="openstack/dnsmasq-dns-f64d5748f-cjt2l" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.058365 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d504ddc2-fb6b-4307-9380-3c4fd24ca8db-dns-svc\") pod \"dnsmasq-dns-f64d5748f-cjt2l\" (UID: \"d504ddc2-fb6b-4307-9380-3c4fd24ca8db\") " pod="openstack/dnsmasq-dns-f64d5748f-cjt2l" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.058381 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d504ddc2-fb6b-4307-9380-3c4fd24ca8db-ovsdbserver-nb\") pod \"dnsmasq-dns-f64d5748f-cjt2l\" (UID: \"d504ddc2-fb6b-4307-9380-3c4fd24ca8db\") " pod="openstack/dnsmasq-dns-f64d5748f-cjt2l" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.058407 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d504ddc2-fb6b-4307-9380-3c4fd24ca8db-ovsdbserver-sb\") pod \"dnsmasq-dns-f64d5748f-cjt2l\" (UID: \"d504ddc2-fb6b-4307-9380-3c4fd24ca8db\") " pod="openstack/dnsmasq-dns-f64d5748f-cjt2l" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.058467 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9a7d7f4-a682-4e90-99e4-b3b2afe2c718-scripts\") pod \"cinder-api-0\" (UID: \"b9a7d7f4-a682-4e90-99e4-b3b2afe2c718\") " pod="openstack/cinder-api-0" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.058494 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6zkr\" (UniqueName: \"kubernetes.io/projected/d504ddc2-fb6b-4307-9380-3c4fd24ca8db-kube-api-access-z6zkr\") pod \"dnsmasq-dns-f64d5748f-cjt2l\" (UID: \"d504ddc2-fb6b-4307-9380-3c4fd24ca8db\") " pod="openstack/dnsmasq-dns-f64d5748f-cjt2l" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.058514 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9a7d7f4-a682-4e90-99e4-b3b2afe2c718-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"b9a7d7f4-a682-4e90-99e4-b3b2afe2c718\") " pod="openstack/cinder-api-0" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.058530 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9a7d7f4-a682-4e90-99e4-b3b2afe2c718-config-data\") pod \"cinder-api-0\" (UID: \"b9a7d7f4-a682-4e90-99e4-b3b2afe2c718\") " pod="openstack/cinder-api-0" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.058550 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b9a7d7f4-a682-4e90-99e4-b3b2afe2c718-logs\") pod \"cinder-api-0\" (UID: \"b9a7d7f4-a682-4e90-99e4-b3b2afe2c718\") " pod="openstack/cinder-api-0" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.062213 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-9b48ddc7c-fn75z"] Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.064229 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-9b48ddc7c-fn75z" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.076550 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.076942 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.077299 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-qmv66" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.079673 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.093457 4873 scope.go:117] "RemoveContainer" containerID="90f9025bf705ac3103b5bb02995016ce74415078a98c44fd7a6326f903c2b040" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.100990 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-9b48ddc7c-fn75z"] Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.161633 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d0787b5-0da8-493b-b55d-b6e36a759f7f-config-data\") pod \"barbican-worker-9b48ddc7c-fn75z\" (UID: \"4d0787b5-0da8-493b-b55d-b6e36a759f7f\") " pod="openstack/barbican-worker-9b48ddc7c-fn75z" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.161688 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d0787b5-0da8-493b-b55d-b6e36a759f7f-combined-ca-bundle\") pod \"barbican-worker-9b48ddc7c-fn75z\" (UID: \"4d0787b5-0da8-493b-b55d-b6e36a759f7f\") " pod="openstack/barbican-worker-9b48ddc7c-fn75z" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.161735 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9a7d7f4-a682-4e90-99e4-b3b2afe2c718-scripts\") pod \"cinder-api-0\" (UID: \"b9a7d7f4-a682-4e90-99e4-b3b2afe2c718\") " pod="openstack/cinder-api-0" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.161772 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6zkr\" (UniqueName: \"kubernetes.io/projected/d504ddc2-fb6b-4307-9380-3c4fd24ca8db-kube-api-access-z6zkr\") pod \"dnsmasq-dns-f64d5748f-cjt2l\" (UID: \"d504ddc2-fb6b-4307-9380-3c4fd24ca8db\") " pod="openstack/dnsmasq-dns-f64d5748f-cjt2l" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.161794 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9a7d7f4-a682-4e90-99e4-b3b2afe2c718-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"b9a7d7f4-a682-4e90-99e4-b3b2afe2c718\") " pod="openstack/cinder-api-0" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.162287 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b9a7d7f4-a682-4e90-99e4-b3b2afe2c718-logs\") pod \"cinder-api-0\" (UID: \"b9a7d7f4-a682-4e90-99e4-b3b2afe2c718\") " pod="openstack/cinder-api-0" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.162320 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9a7d7f4-a682-4e90-99e4-b3b2afe2c718-config-data\") pod \"cinder-api-0\" (UID: \"b9a7d7f4-a682-4e90-99e4-b3b2afe2c718\") " pod="openstack/cinder-api-0" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.162358 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtdwz\" (UniqueName: \"kubernetes.io/projected/4d0787b5-0da8-493b-b55d-b6e36a759f7f-kube-api-access-gtdwz\") pod \"barbican-worker-9b48ddc7c-fn75z\" (UID: \"4d0787b5-0da8-493b-b55d-b6e36a759f7f\") " pod="openstack/barbican-worker-9b48ddc7c-fn75z" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.162398 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b9a7d7f4-a682-4e90-99e4-b3b2afe2c718-etc-machine-id\") pod \"cinder-api-0\" (UID: \"b9a7d7f4-a682-4e90-99e4-b3b2afe2c718\") " pod="openstack/cinder-api-0" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.162416 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swj9t\" (UniqueName: \"kubernetes.io/projected/b9a7d7f4-a682-4e90-99e4-b3b2afe2c718-kube-api-access-swj9t\") pod \"cinder-api-0\" (UID: \"b9a7d7f4-a682-4e90-99e4-b3b2afe2c718\") " pod="openstack/cinder-api-0" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.162455 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b9a7d7f4-a682-4e90-99e4-b3b2afe2c718-config-data-custom\") pod \"cinder-api-0\" (UID: \"b9a7d7f4-a682-4e90-99e4-b3b2afe2c718\") " pod="openstack/cinder-api-0" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.162477 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d504ddc2-fb6b-4307-9380-3c4fd24ca8db-config\") pod \"dnsmasq-dns-f64d5748f-cjt2l\" (UID: \"d504ddc2-fb6b-4307-9380-3c4fd24ca8db\") " pod="openstack/dnsmasq-dns-f64d5748f-cjt2l" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.162508 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4d0787b5-0da8-493b-b55d-b6e36a759f7f-config-data-custom\") pod \"barbican-worker-9b48ddc7c-fn75z\" (UID: \"4d0787b5-0da8-493b-b55d-b6e36a759f7f\") " pod="openstack/barbican-worker-9b48ddc7c-fn75z" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.162545 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d504ddc2-fb6b-4307-9380-3c4fd24ca8db-dns-svc\") pod \"dnsmasq-dns-f64d5748f-cjt2l\" (UID: \"d504ddc2-fb6b-4307-9380-3c4fd24ca8db\") " pod="openstack/dnsmasq-dns-f64d5748f-cjt2l" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.162567 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d504ddc2-fb6b-4307-9380-3c4fd24ca8db-ovsdbserver-nb\") pod \"dnsmasq-dns-f64d5748f-cjt2l\" (UID: \"d504ddc2-fb6b-4307-9380-3c4fd24ca8db\") " pod="openstack/dnsmasq-dns-f64d5748f-cjt2l" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.162594 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d504ddc2-fb6b-4307-9380-3c4fd24ca8db-ovsdbserver-sb\") pod \"dnsmasq-dns-f64d5748f-cjt2l\" (UID: \"d504ddc2-fb6b-4307-9380-3c4fd24ca8db\") " pod="openstack/dnsmasq-dns-f64d5748f-cjt2l" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.162623 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4d0787b5-0da8-493b-b55d-b6e36a759f7f-logs\") pod \"barbican-worker-9b48ddc7c-fn75z\" (UID: \"4d0787b5-0da8-493b-b55d-b6e36a759f7f\") " pod="openstack/barbican-worker-9b48ddc7c-fn75z" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.163377 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b9a7d7f4-a682-4e90-99e4-b3b2afe2c718-etc-machine-id\") pod \"cinder-api-0\" (UID: \"b9a7d7f4-a682-4e90-99e4-b3b2afe2c718\") " pod="openstack/cinder-api-0" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.166585 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d504ddc2-fb6b-4307-9380-3c4fd24ca8db-config\") pod \"dnsmasq-dns-f64d5748f-cjt2l\" (UID: \"d504ddc2-fb6b-4307-9380-3c4fd24ca8db\") " pod="openstack/dnsmasq-dns-f64d5748f-cjt2l" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.167946 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b9a7d7f4-a682-4e90-99e4-b3b2afe2c718-logs\") pod \"cinder-api-0\" (UID: \"b9a7d7f4-a682-4e90-99e4-b3b2afe2c718\") " pod="openstack/cinder-api-0" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.169037 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d504ddc2-fb6b-4307-9380-3c4fd24ca8db-dns-svc\") pod \"dnsmasq-dns-f64d5748f-cjt2l\" (UID: \"d504ddc2-fb6b-4307-9380-3c4fd24ca8db\") " pod="openstack/dnsmasq-dns-f64d5748f-cjt2l" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.179450 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d504ddc2-fb6b-4307-9380-3c4fd24ca8db-ovsdbserver-sb\") pod \"dnsmasq-dns-f64d5748f-cjt2l\" (UID: \"d504ddc2-fb6b-4307-9380-3c4fd24ca8db\") " pod="openstack/dnsmasq-dns-f64d5748f-cjt2l" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.179823 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d504ddc2-fb6b-4307-9380-3c4fd24ca8db-ovsdbserver-nb\") pod \"dnsmasq-dns-f64d5748f-cjt2l\" (UID: \"d504ddc2-fb6b-4307-9380-3c4fd24ca8db\") " pod="openstack/dnsmasq-dns-f64d5748f-cjt2l" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.180888 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-6674fb675b-58jdf"] Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.182739 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6674fb675b-58jdf" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.196891 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b9a7d7f4-a682-4e90-99e4-b3b2afe2c718-config-data-custom\") pod \"cinder-api-0\" (UID: \"b9a7d7f4-a682-4e90-99e4-b3b2afe2c718\") " pod="openstack/cinder-api-0" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.196983 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7987f74bbc-6rfmd"] Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.201031 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.204956 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9a7d7f4-a682-4e90-99e4-b3b2afe2c718-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"b9a7d7f4-a682-4e90-99e4-b3b2afe2c718\") " pod="openstack/cinder-api-0" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.207883 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9a7d7f4-a682-4e90-99e4-b3b2afe2c718-config-data\") pod \"cinder-api-0\" (UID: \"b9a7d7f4-a682-4e90-99e4-b3b2afe2c718\") " pod="openstack/cinder-api-0" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.212427 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6zkr\" (UniqueName: \"kubernetes.io/projected/d504ddc2-fb6b-4307-9380-3c4fd24ca8db-kube-api-access-z6zkr\") pod \"dnsmasq-dns-f64d5748f-cjt2l\" (UID: \"d504ddc2-fb6b-4307-9380-3c4fd24ca8db\") " pod="openstack/dnsmasq-dns-f64d5748f-cjt2l" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.218437 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7987f74bbc-6rfmd"] Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.219183 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9a7d7f4-a682-4e90-99e4-b3b2afe2c718-scripts\") pod \"cinder-api-0\" (UID: \"b9a7d7f4-a682-4e90-99e4-b3b2afe2c718\") " pod="openstack/cinder-api-0" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.227755 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swj9t\" (UniqueName: \"kubernetes.io/projected/b9a7d7f4-a682-4e90-99e4-b3b2afe2c718-kube-api-access-swj9t\") pod \"cinder-api-0\" (UID: \"b9a7d7f4-a682-4e90-99e4-b3b2afe2c718\") " pod="openstack/cinder-api-0" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.227849 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6674fb675b-58jdf"] Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.249454 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f64d5748f-cjt2l"] Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.250329 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f64d5748f-cjt2l" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.276626 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4d0787b5-0da8-493b-b55d-b6e36a759f7f-config-data-custom\") pod \"barbican-worker-9b48ddc7c-fn75z\" (UID: \"4d0787b5-0da8-493b-b55d-b6e36a759f7f\") " pod="openstack/barbican-worker-9b48ddc7c-fn75z" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.276725 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/99f7daba-8483-4494-b783-9628f828ea49-config-data-custom\") pod \"barbican-keystone-listener-6674fb675b-58jdf\" (UID: \"99f7daba-8483-4494-b783-9628f828ea49\") " pod="openstack/barbican-keystone-listener-6674fb675b-58jdf" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.276778 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4d0787b5-0da8-493b-b55d-b6e36a759f7f-logs\") pod \"barbican-worker-9b48ddc7c-fn75z\" (UID: \"4d0787b5-0da8-493b-b55d-b6e36a759f7f\") " pod="openstack/barbican-worker-9b48ddc7c-fn75z" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.276819 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d0787b5-0da8-493b-b55d-b6e36a759f7f-config-data\") pod \"barbican-worker-9b48ddc7c-fn75z\" (UID: \"4d0787b5-0da8-493b-b55d-b6e36a759f7f\") " pod="openstack/barbican-worker-9b48ddc7c-fn75z" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.276845 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d0787b5-0da8-493b-b55d-b6e36a759f7f-combined-ca-bundle\") pod \"barbican-worker-9b48ddc7c-fn75z\" (UID: \"4d0787b5-0da8-493b-b55d-b6e36a759f7f\") " pod="openstack/barbican-worker-9b48ddc7c-fn75z" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.276880 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99f7daba-8483-4494-b783-9628f828ea49-combined-ca-bundle\") pod \"barbican-keystone-listener-6674fb675b-58jdf\" (UID: \"99f7daba-8483-4494-b783-9628f828ea49\") " pod="openstack/barbican-keystone-listener-6674fb675b-58jdf" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.276922 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtdwz\" (UniqueName: \"kubernetes.io/projected/4d0787b5-0da8-493b-b55d-b6e36a759f7f-kube-api-access-gtdwz\") pod \"barbican-worker-9b48ddc7c-fn75z\" (UID: \"4d0787b5-0da8-493b-b55d-b6e36a759f7f\") " pod="openstack/barbican-worker-9b48ddc7c-fn75z" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.276965 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6q79j\" (UniqueName: \"kubernetes.io/projected/99f7daba-8483-4494-b783-9628f828ea49-kube-api-access-6q79j\") pod \"barbican-keystone-listener-6674fb675b-58jdf\" (UID: \"99f7daba-8483-4494-b783-9628f828ea49\") " pod="openstack/barbican-keystone-listener-6674fb675b-58jdf" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.276990 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99f7daba-8483-4494-b783-9628f828ea49-logs\") pod \"barbican-keystone-listener-6674fb675b-58jdf\" (UID: \"99f7daba-8483-4494-b783-9628f828ea49\") " pod="openstack/barbican-keystone-listener-6674fb675b-58jdf" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.277031 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99f7daba-8483-4494-b783-9628f828ea49-config-data\") pod \"barbican-keystone-listener-6674fb675b-58jdf\" (UID: \"99f7daba-8483-4494-b783-9628f828ea49\") " pod="openstack/barbican-keystone-listener-6674fb675b-58jdf" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.286826 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4d0787b5-0da8-493b-b55d-b6e36a759f7f-logs\") pod \"barbican-worker-9b48ddc7c-fn75z\" (UID: \"4d0787b5-0da8-493b-b55d-b6e36a759f7f\") " pod="openstack/barbican-worker-9b48ddc7c-fn75z" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.290240 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d0787b5-0da8-493b-b55d-b6e36a759f7f-combined-ca-bundle\") pod \"barbican-worker-9b48ddc7c-fn75z\" (UID: \"4d0787b5-0da8-493b-b55d-b6e36a759f7f\") " pod="openstack/barbican-worker-9b48ddc7c-fn75z" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.290295 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4d0787b5-0da8-493b-b55d-b6e36a759f7f-config-data-custom\") pod \"barbican-worker-9b48ddc7c-fn75z\" (UID: \"4d0787b5-0da8-493b-b55d-b6e36a759f7f\") " pod="openstack/barbican-worker-9b48ddc7c-fn75z" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.290701 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d0787b5-0da8-493b-b55d-b6e36a759f7f-config-data\") pod \"barbican-worker-9b48ddc7c-fn75z\" (UID: \"4d0787b5-0da8-493b-b55d-b6e36a759f7f\") " pod="openstack/barbican-worker-9b48ddc7c-fn75z" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.403881 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/99f7daba-8483-4494-b783-9628f828ea49-config-data-custom\") pod \"barbican-keystone-listener-6674fb675b-58jdf\" (UID: \"99f7daba-8483-4494-b783-9628f828ea49\") " pod="openstack/barbican-keystone-listener-6674fb675b-58jdf" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.403980 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99f7daba-8483-4494-b783-9628f828ea49-combined-ca-bundle\") pod \"barbican-keystone-listener-6674fb675b-58jdf\" (UID: \"99f7daba-8483-4494-b783-9628f828ea49\") " pod="openstack/barbican-keystone-listener-6674fb675b-58jdf" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.404084 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6q79j\" (UniqueName: \"kubernetes.io/projected/99f7daba-8483-4494-b783-9628f828ea49-kube-api-access-6q79j\") pod \"barbican-keystone-listener-6674fb675b-58jdf\" (UID: \"99f7daba-8483-4494-b783-9628f828ea49\") " pod="openstack/barbican-keystone-listener-6674fb675b-58jdf" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.404111 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99f7daba-8483-4494-b783-9628f828ea49-logs\") pod \"barbican-keystone-listener-6674fb675b-58jdf\" (UID: \"99f7daba-8483-4494-b783-9628f828ea49\") " pod="openstack/barbican-keystone-listener-6674fb675b-58jdf" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.404129 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99f7daba-8483-4494-b783-9628f828ea49-config-data\") pod \"barbican-keystone-listener-6674fb675b-58jdf\" (UID: \"99f7daba-8483-4494-b783-9628f828ea49\") " pod="openstack/barbican-keystone-listener-6674fb675b-58jdf" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.408576 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99f7daba-8483-4494-b783-9628f828ea49-logs\") pod \"barbican-keystone-listener-6674fb675b-58jdf\" (UID: \"99f7daba-8483-4494-b783-9628f828ea49\") " pod="openstack/barbican-keystone-listener-6674fb675b-58jdf" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.412526 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.418034 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtdwz\" (UniqueName: \"kubernetes.io/projected/4d0787b5-0da8-493b-b55d-b6e36a759f7f-kube-api-access-gtdwz\") pod \"barbican-worker-9b48ddc7c-fn75z\" (UID: \"4d0787b5-0da8-493b-b55d-b6e36a759f7f\") " pod="openstack/barbican-worker-9b48ddc7c-fn75z" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.420652 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/99f7daba-8483-4494-b783-9628f828ea49-config-data-custom\") pod \"barbican-keystone-listener-6674fb675b-58jdf\" (UID: \"99f7daba-8483-4494-b783-9628f828ea49\") " pod="openstack/barbican-keystone-listener-6674fb675b-58jdf" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.420838 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99f7daba-8483-4494-b783-9628f828ea49-config-data\") pod \"barbican-keystone-listener-6674fb675b-58jdf\" (UID: \"99f7daba-8483-4494-b783-9628f828ea49\") " pod="openstack/barbican-keystone-listener-6674fb675b-58jdf" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.428072 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d97fcdd8f-qmg5n"] Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.428173 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99f7daba-8483-4494-b783-9628f828ea49-combined-ca-bundle\") pod \"barbican-keystone-listener-6674fb675b-58jdf\" (UID: \"99f7daba-8483-4494-b783-9628f828ea49\") " pod="openstack/barbican-keystone-listener-6674fb675b-58jdf" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.432196 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d97fcdd8f-qmg5n" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.446625 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6q79j\" (UniqueName: \"kubernetes.io/projected/99f7daba-8483-4494-b783-9628f828ea49-kube-api-access-6q79j\") pod \"barbican-keystone-listener-6674fb675b-58jdf\" (UID: \"99f7daba-8483-4494-b783-9628f828ea49\") " pod="openstack/barbican-keystone-listener-6674fb675b-58jdf" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.475857 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d97fcdd8f-qmg5n"] Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.590314 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-64c5c798d6-h87s6"] Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.594114 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-64c5c798d6-h87s6" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.594849 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-9b48ddc7c-fn75z" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.596762 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.619135 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-64c5c798d6-h87s6"] Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.634663 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mj4f8\" (UniqueName: \"kubernetes.io/projected/c10e1747-9f3e-4fc3-bd55-b83c7a93ea12-kube-api-access-mj4f8\") pod \"barbican-api-64c5c798d6-h87s6\" (UID: \"c10e1747-9f3e-4fc3-bd55-b83c7a93ea12\") " pod="openstack/barbican-api-64c5c798d6-h87s6" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.634742 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c10e1747-9f3e-4fc3-bd55-b83c7a93ea12-config-data\") pod \"barbican-api-64c5c798d6-h87s6\" (UID: \"c10e1747-9f3e-4fc3-bd55-b83c7a93ea12\") " pod="openstack/barbican-api-64c5c798d6-h87s6" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.634781 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r67jf\" (UniqueName: \"kubernetes.io/projected/a6a491d2-c38c-41ea-a1bd-ca6f6b276bde-kube-api-access-r67jf\") pod \"dnsmasq-dns-6d97fcdd8f-qmg5n\" (UID: \"a6a491d2-c38c-41ea-a1bd-ca6f6b276bde\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-qmg5n" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.638881 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c10e1747-9f3e-4fc3-bd55-b83c7a93ea12-config-data-custom\") pod \"barbican-api-64c5c798d6-h87s6\" (UID: \"c10e1747-9f3e-4fc3-bd55-b83c7a93ea12\") " pod="openstack/barbican-api-64c5c798d6-h87s6" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.639813 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6a491d2-c38c-41ea-a1bd-ca6f6b276bde-config\") pod \"dnsmasq-dns-6d97fcdd8f-qmg5n\" (UID: \"a6a491d2-c38c-41ea-a1bd-ca6f6b276bde\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-qmg5n" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.639862 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a6a491d2-c38c-41ea-a1bd-ca6f6b276bde-dns-svc\") pod \"dnsmasq-dns-6d97fcdd8f-qmg5n\" (UID: \"a6a491d2-c38c-41ea-a1bd-ca6f6b276bde\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-qmg5n" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.639956 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c10e1747-9f3e-4fc3-bd55-b83c7a93ea12-combined-ca-bundle\") pod \"barbican-api-64c5c798d6-h87s6\" (UID: \"c10e1747-9f3e-4fc3-bd55-b83c7a93ea12\") " pod="openstack/barbican-api-64c5c798d6-h87s6" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.640050 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a6a491d2-c38c-41ea-a1bd-ca6f6b276bde-ovsdbserver-sb\") pod \"dnsmasq-dns-6d97fcdd8f-qmg5n\" (UID: \"a6a491d2-c38c-41ea-a1bd-ca6f6b276bde\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-qmg5n" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.640109 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c10e1747-9f3e-4fc3-bd55-b83c7a93ea12-logs\") pod \"barbican-api-64c5c798d6-h87s6\" (UID: \"c10e1747-9f3e-4fc3-bd55-b83c7a93ea12\") " pod="openstack/barbican-api-64c5c798d6-h87s6" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.640166 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a6a491d2-c38c-41ea-a1bd-ca6f6b276bde-ovsdbserver-nb\") pod \"dnsmasq-dns-6d97fcdd8f-qmg5n\" (UID: \"a6a491d2-c38c-41ea-a1bd-ca6f6b276bde\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-qmg5n" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.715852 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6674fb675b-58jdf" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.742854 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c10e1747-9f3e-4fc3-bd55-b83c7a93ea12-config-data\") pod \"barbican-api-64c5c798d6-h87s6\" (UID: \"c10e1747-9f3e-4fc3-bd55-b83c7a93ea12\") " pod="openstack/barbican-api-64c5c798d6-h87s6" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.742931 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r67jf\" (UniqueName: \"kubernetes.io/projected/a6a491d2-c38c-41ea-a1bd-ca6f6b276bde-kube-api-access-r67jf\") pod \"dnsmasq-dns-6d97fcdd8f-qmg5n\" (UID: \"a6a491d2-c38c-41ea-a1bd-ca6f6b276bde\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-qmg5n" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.743052 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c10e1747-9f3e-4fc3-bd55-b83c7a93ea12-config-data-custom\") pod \"barbican-api-64c5c798d6-h87s6\" (UID: \"c10e1747-9f3e-4fc3-bd55-b83c7a93ea12\") " pod="openstack/barbican-api-64c5c798d6-h87s6" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.743137 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6a491d2-c38c-41ea-a1bd-ca6f6b276bde-config\") pod \"dnsmasq-dns-6d97fcdd8f-qmg5n\" (UID: \"a6a491d2-c38c-41ea-a1bd-ca6f6b276bde\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-qmg5n" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.743178 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a6a491d2-c38c-41ea-a1bd-ca6f6b276bde-dns-svc\") pod \"dnsmasq-dns-6d97fcdd8f-qmg5n\" (UID: \"a6a491d2-c38c-41ea-a1bd-ca6f6b276bde\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-qmg5n" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.743251 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c10e1747-9f3e-4fc3-bd55-b83c7a93ea12-combined-ca-bundle\") pod \"barbican-api-64c5c798d6-h87s6\" (UID: \"c10e1747-9f3e-4fc3-bd55-b83c7a93ea12\") " pod="openstack/barbican-api-64c5c798d6-h87s6" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.743297 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a6a491d2-c38c-41ea-a1bd-ca6f6b276bde-ovsdbserver-sb\") pod \"dnsmasq-dns-6d97fcdd8f-qmg5n\" (UID: \"a6a491d2-c38c-41ea-a1bd-ca6f6b276bde\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-qmg5n" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.743352 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c10e1747-9f3e-4fc3-bd55-b83c7a93ea12-logs\") pod \"barbican-api-64c5c798d6-h87s6\" (UID: \"c10e1747-9f3e-4fc3-bd55-b83c7a93ea12\") " pod="openstack/barbican-api-64c5c798d6-h87s6" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.743416 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a6a491d2-c38c-41ea-a1bd-ca6f6b276bde-ovsdbserver-nb\") pod \"dnsmasq-dns-6d97fcdd8f-qmg5n\" (UID: \"a6a491d2-c38c-41ea-a1bd-ca6f6b276bde\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-qmg5n" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.743446 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mj4f8\" (UniqueName: \"kubernetes.io/projected/c10e1747-9f3e-4fc3-bd55-b83c7a93ea12-kube-api-access-mj4f8\") pod \"barbican-api-64c5c798d6-h87s6\" (UID: \"c10e1747-9f3e-4fc3-bd55-b83c7a93ea12\") " pod="openstack/barbican-api-64c5c798d6-h87s6" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.745084 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a6a491d2-c38c-41ea-a1bd-ca6f6b276bde-ovsdbserver-sb\") pod \"dnsmasq-dns-6d97fcdd8f-qmg5n\" (UID: \"a6a491d2-c38c-41ea-a1bd-ca6f6b276bde\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-qmg5n" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.745080 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6a491d2-c38c-41ea-a1bd-ca6f6b276bde-config\") pod \"dnsmasq-dns-6d97fcdd8f-qmg5n\" (UID: \"a6a491d2-c38c-41ea-a1bd-ca6f6b276bde\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-qmg5n" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.745415 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c10e1747-9f3e-4fc3-bd55-b83c7a93ea12-logs\") pod \"barbican-api-64c5c798d6-h87s6\" (UID: \"c10e1747-9f3e-4fc3-bd55-b83c7a93ea12\") " pod="openstack/barbican-api-64c5c798d6-h87s6" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.746098 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a6a491d2-c38c-41ea-a1bd-ca6f6b276bde-ovsdbserver-nb\") pod \"dnsmasq-dns-6d97fcdd8f-qmg5n\" (UID: \"a6a491d2-c38c-41ea-a1bd-ca6f6b276bde\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-qmg5n" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.752665 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a6a491d2-c38c-41ea-a1bd-ca6f6b276bde-dns-svc\") pod \"dnsmasq-dns-6d97fcdd8f-qmg5n\" (UID: \"a6a491d2-c38c-41ea-a1bd-ca6f6b276bde\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-qmg5n" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.755912 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c10e1747-9f3e-4fc3-bd55-b83c7a93ea12-combined-ca-bundle\") pod \"barbican-api-64c5c798d6-h87s6\" (UID: \"c10e1747-9f3e-4fc3-bd55-b83c7a93ea12\") " pod="openstack/barbican-api-64c5c798d6-h87s6" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.760785 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c10e1747-9f3e-4fc3-bd55-b83c7a93ea12-config-data-custom\") pod \"barbican-api-64c5c798d6-h87s6\" (UID: \"c10e1747-9f3e-4fc3-bd55-b83c7a93ea12\") " pod="openstack/barbican-api-64c5c798d6-h87s6" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.763701 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r67jf\" (UniqueName: \"kubernetes.io/projected/a6a491d2-c38c-41ea-a1bd-ca6f6b276bde-kube-api-access-r67jf\") pod \"dnsmasq-dns-6d97fcdd8f-qmg5n\" (UID: \"a6a491d2-c38c-41ea-a1bd-ca6f6b276bde\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-qmg5n" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.764679 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mj4f8\" (UniqueName: \"kubernetes.io/projected/c10e1747-9f3e-4fc3-bd55-b83c7a93ea12-kube-api-access-mj4f8\") pod \"barbican-api-64c5c798d6-h87s6\" (UID: \"c10e1747-9f3e-4fc3-bd55-b83c7a93ea12\") " pod="openstack/barbican-api-64c5c798d6-h87s6" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.768419 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c10e1747-9f3e-4fc3-bd55-b83c7a93ea12-config-data\") pod \"barbican-api-64c5c798d6-h87s6\" (UID: \"c10e1747-9f3e-4fc3-bd55-b83c7a93ea12\") " pod="openstack/barbican-api-64c5c798d6-h87s6" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.789859 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d97fcdd8f-qmg5n" Dec 01 09:00:09 crc kubenswrapper[4873]: I1201 09:00:09.982673 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-64c5c798d6-h87s6" Dec 01 09:00:10 crc kubenswrapper[4873]: I1201 09:00:10.071292 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 09:00:10 crc kubenswrapper[4873]: I1201 09:00:10.078207 4873 generic.go:334] "Generic (PLEG): container finished" podID="e53689a0-5d92-4e62-bb63-3a56af762a59" containerID="068b339c05967459adb9767502db98e3ed7f8097f9d5a942acefdc2ee1b56026" exitCode=0 Dec 01 09:00:10 crc kubenswrapper[4873]: I1201 09:00:10.078257 4873 generic.go:334] "Generic (PLEG): container finished" podID="e53689a0-5d92-4e62-bb63-3a56af762a59" containerID="b33b0f231101926f73eefa68214880e4c14df8323cb5ce1c1750a3ab4b4e019d" exitCode=2 Dec 01 09:00:10 crc kubenswrapper[4873]: I1201 09:00:10.078345 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e53689a0-5d92-4e62-bb63-3a56af762a59","Type":"ContainerDied","Data":"068b339c05967459adb9767502db98e3ed7f8097f9d5a942acefdc2ee1b56026"} Dec 01 09:00:10 crc kubenswrapper[4873]: I1201 09:00:10.078378 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e53689a0-5d92-4e62-bb63-3a56af762a59","Type":"ContainerDied","Data":"b33b0f231101926f73eefa68214880e4c14df8323cb5ce1c1750a3ab4b4e019d"} Dec 01 09:00:10 crc kubenswrapper[4873]: I1201 09:00:10.170628 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 01 09:00:10 crc kubenswrapper[4873]: I1201 09:00:10.189960 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f64d5748f-cjt2l"] Dec 01 09:00:10 crc kubenswrapper[4873]: I1201 09:00:10.447256 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f027d3a-e35f-418b-bc8b-508a637f6698" path="/var/lib/kubelet/pods/4f027d3a-e35f-418b-bc8b-508a637f6698/volumes" Dec 01 09:00:10 crc kubenswrapper[4873]: I1201 09:00:10.517478 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-9b48ddc7c-fn75z"] Dec 01 09:00:10 crc kubenswrapper[4873]: I1201 09:00:10.528064 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6674fb675b-58jdf"] Dec 01 09:00:10 crc kubenswrapper[4873]: W1201 09:00:10.529971 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d0787b5_0da8_493b_b55d_b6e36a759f7f.slice/crio-9754c5cc9bc2df74f8720a88c0fa90e9c97296aaf76e04af4340faa93af133a1 WatchSource:0}: Error finding container 9754c5cc9bc2df74f8720a88c0fa90e9c97296aaf76e04af4340faa93af133a1: Status 404 returned error can't find the container with id 9754c5cc9bc2df74f8720a88c0fa90e9c97296aaf76e04af4340faa93af133a1 Dec 01 09:00:10 crc kubenswrapper[4873]: W1201 09:00:10.531698 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99f7daba_8483_4494_b783_9628f828ea49.slice/crio-660fcb69dc4c53a9d3ad7cf46217d117200557eef8a3f92cd2fffa2aba53f326 WatchSource:0}: Error finding container 660fcb69dc4c53a9d3ad7cf46217d117200557eef8a3f92cd2fffa2aba53f326: Status 404 returned error can't find the container with id 660fcb69dc4c53a9d3ad7cf46217d117200557eef8a3f92cd2fffa2aba53f326 Dec 01 09:00:10 crc kubenswrapper[4873]: I1201 09:00:10.652069 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d97fcdd8f-qmg5n"] Dec 01 09:00:10 crc kubenswrapper[4873]: W1201 09:00:10.658383 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6a491d2_c38c_41ea_a1bd_ca6f6b276bde.slice/crio-201142b7bf24cbd8e870f489422fb63ddf91d8caccf362e633e31a0ce5abe193 WatchSource:0}: Error finding container 201142b7bf24cbd8e870f489422fb63ddf91d8caccf362e633e31a0ce5abe193: Status 404 returned error can't find the container with id 201142b7bf24cbd8e870f489422fb63ddf91d8caccf362e633e31a0ce5abe193 Dec 01 09:00:10 crc kubenswrapper[4873]: I1201 09:00:10.746064 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-64c5c798d6-h87s6"] Dec 01 09:00:10 crc kubenswrapper[4873]: W1201 09:00:10.755823 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc10e1747_9f3e_4fc3_bd55_b83c7a93ea12.slice/crio-721d8e182e7f528d1281b938b4c11f54863a9778f23f7c3b9f7de84e7e36cb59 WatchSource:0}: Error finding container 721d8e182e7f528d1281b938b4c11f54863a9778f23f7c3b9f7de84e7e36cb59: Status 404 returned error can't find the container with id 721d8e182e7f528d1281b938b4c11f54863a9778f23f7c3b9f7de84e7e36cb59 Dec 01 09:00:11 crc kubenswrapper[4873]: I1201 09:00:11.176220 4873 generic.go:334] "Generic (PLEG): container finished" podID="d504ddc2-fb6b-4307-9380-3c4fd24ca8db" containerID="48e6422ff5a86c5a3fd6771d60481e6825da808c2a512d77ca2c812a65aee862" exitCode=0 Dec 01 09:00:11 crc kubenswrapper[4873]: I1201 09:00:11.176605 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f64d5748f-cjt2l" event={"ID":"d504ddc2-fb6b-4307-9380-3c4fd24ca8db","Type":"ContainerDied","Data":"48e6422ff5a86c5a3fd6771d60481e6825da808c2a512d77ca2c812a65aee862"} Dec 01 09:00:11 crc kubenswrapper[4873]: I1201 09:00:11.176646 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f64d5748f-cjt2l" event={"ID":"d504ddc2-fb6b-4307-9380-3c4fd24ca8db","Type":"ContainerStarted","Data":"4f608b453c44eed55aa497a873194c0a0b0dd82ecdbec86335dce3b27f4a6182"} Dec 01 09:00:11 crc kubenswrapper[4873]: I1201 09:00:11.192389 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b9a7d7f4-a682-4e90-99e4-b3b2afe2c718","Type":"ContainerStarted","Data":"818f3fa6d352a247655200dfe50d9ebb7fb5071a15a3cfcb04b3cddad6efedbe"} Dec 01 09:00:11 crc kubenswrapper[4873]: I1201 09:00:11.203532 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"11731b74-fe99-40c4-9445-f86102242d05","Type":"ContainerStarted","Data":"8358bfa8cb72a469f44ad316593ed5b68a569b1833abfacd2c88aeba8cd0800c"} Dec 01 09:00:11 crc kubenswrapper[4873]: I1201 09:00:11.215002 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-64c5c798d6-h87s6" event={"ID":"c10e1747-9f3e-4fc3-bd55-b83c7a93ea12","Type":"ContainerStarted","Data":"9e26d54a7fa344ea38e08b4d1b1bfb07202870fd5bd4d23df69645874876f965"} Dec 01 09:00:11 crc kubenswrapper[4873]: I1201 09:00:11.215090 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-64c5c798d6-h87s6" event={"ID":"c10e1747-9f3e-4fc3-bd55-b83c7a93ea12","Type":"ContainerStarted","Data":"721d8e182e7f528d1281b938b4c11f54863a9778f23f7c3b9f7de84e7e36cb59"} Dec 01 09:00:11 crc kubenswrapper[4873]: I1201 09:00:11.218054 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 01 09:00:11 crc kubenswrapper[4873]: I1201 09:00:11.218863 4873 generic.go:334] "Generic (PLEG): container finished" podID="a6a491d2-c38c-41ea-a1bd-ca6f6b276bde" containerID="5ee407ae1406fab0318c27c00c0b6d42c4759ae2fcd6462713d61cd472c55418" exitCode=0 Dec 01 09:00:11 crc kubenswrapper[4873]: I1201 09:00:11.218952 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d97fcdd8f-qmg5n" event={"ID":"a6a491d2-c38c-41ea-a1bd-ca6f6b276bde","Type":"ContainerDied","Data":"5ee407ae1406fab0318c27c00c0b6d42c4759ae2fcd6462713d61cd472c55418"} Dec 01 09:00:11 crc kubenswrapper[4873]: I1201 09:00:11.218984 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d97fcdd8f-qmg5n" event={"ID":"a6a491d2-c38c-41ea-a1bd-ca6f6b276bde","Type":"ContainerStarted","Data":"201142b7bf24cbd8e870f489422fb63ddf91d8caccf362e633e31a0ce5abe193"} Dec 01 09:00:11 crc kubenswrapper[4873]: I1201 09:00:11.224683 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-9b48ddc7c-fn75z" event={"ID":"4d0787b5-0da8-493b-b55d-b6e36a759f7f","Type":"ContainerStarted","Data":"9754c5cc9bc2df74f8720a88c0fa90e9c97296aaf76e04af4340faa93af133a1"} Dec 01 09:00:11 crc kubenswrapper[4873]: I1201 09:00:11.227806 4873 generic.go:334] "Generic (PLEG): container finished" podID="e53689a0-5d92-4e62-bb63-3a56af762a59" containerID="31940bfd155d788c99f80b4389ebba5de7bb5078293bf263879bda05942db53b" exitCode=0 Dec 01 09:00:11 crc kubenswrapper[4873]: I1201 09:00:11.227868 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e53689a0-5d92-4e62-bb63-3a56af762a59","Type":"ContainerDied","Data":"31940bfd155d788c99f80b4389ebba5de7bb5078293bf263879bda05942db53b"} Dec 01 09:00:11 crc kubenswrapper[4873]: I1201 09:00:11.232638 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6674fb675b-58jdf" event={"ID":"99f7daba-8483-4494-b783-9628f828ea49","Type":"ContainerStarted","Data":"660fcb69dc4c53a9d3ad7cf46217d117200557eef8a3f92cd2fffa2aba53f326"} Dec 01 09:00:11 crc kubenswrapper[4873]: I1201 09:00:11.529381 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f64d5748f-cjt2l" Dec 01 09:00:11 crc kubenswrapper[4873]: I1201 09:00:11.662864 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z6zkr\" (UniqueName: \"kubernetes.io/projected/d504ddc2-fb6b-4307-9380-3c4fd24ca8db-kube-api-access-z6zkr\") pod \"d504ddc2-fb6b-4307-9380-3c4fd24ca8db\" (UID: \"d504ddc2-fb6b-4307-9380-3c4fd24ca8db\") " Dec 01 09:00:11 crc kubenswrapper[4873]: I1201 09:00:11.663436 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d504ddc2-fb6b-4307-9380-3c4fd24ca8db-config\") pod \"d504ddc2-fb6b-4307-9380-3c4fd24ca8db\" (UID: \"d504ddc2-fb6b-4307-9380-3c4fd24ca8db\") " Dec 01 09:00:11 crc kubenswrapper[4873]: I1201 09:00:11.663505 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d504ddc2-fb6b-4307-9380-3c4fd24ca8db-ovsdbserver-sb\") pod \"d504ddc2-fb6b-4307-9380-3c4fd24ca8db\" (UID: \"d504ddc2-fb6b-4307-9380-3c4fd24ca8db\") " Dec 01 09:00:11 crc kubenswrapper[4873]: I1201 09:00:11.663711 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d504ddc2-fb6b-4307-9380-3c4fd24ca8db-ovsdbserver-nb\") pod \"d504ddc2-fb6b-4307-9380-3c4fd24ca8db\" (UID: \"d504ddc2-fb6b-4307-9380-3c4fd24ca8db\") " Dec 01 09:00:11 crc kubenswrapper[4873]: I1201 09:00:11.663764 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d504ddc2-fb6b-4307-9380-3c4fd24ca8db-dns-svc\") pod \"d504ddc2-fb6b-4307-9380-3c4fd24ca8db\" (UID: \"d504ddc2-fb6b-4307-9380-3c4fd24ca8db\") " Dec 01 09:00:11 crc kubenswrapper[4873]: I1201 09:00:11.677071 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d504ddc2-fb6b-4307-9380-3c4fd24ca8db-kube-api-access-z6zkr" (OuterVolumeSpecName: "kube-api-access-z6zkr") pod "d504ddc2-fb6b-4307-9380-3c4fd24ca8db" (UID: "d504ddc2-fb6b-4307-9380-3c4fd24ca8db"). InnerVolumeSpecName "kube-api-access-z6zkr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:00:11 crc kubenswrapper[4873]: I1201 09:00:11.700282 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d504ddc2-fb6b-4307-9380-3c4fd24ca8db-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d504ddc2-fb6b-4307-9380-3c4fd24ca8db" (UID: "d504ddc2-fb6b-4307-9380-3c4fd24ca8db"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:00:11 crc kubenswrapper[4873]: I1201 09:00:11.709172 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d504ddc2-fb6b-4307-9380-3c4fd24ca8db-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d504ddc2-fb6b-4307-9380-3c4fd24ca8db" (UID: "d504ddc2-fb6b-4307-9380-3c4fd24ca8db"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:00:11 crc kubenswrapper[4873]: I1201 09:00:11.734301 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d504ddc2-fb6b-4307-9380-3c4fd24ca8db-config" (OuterVolumeSpecName: "config") pod "d504ddc2-fb6b-4307-9380-3c4fd24ca8db" (UID: "d504ddc2-fb6b-4307-9380-3c4fd24ca8db"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:00:11 crc kubenswrapper[4873]: I1201 09:00:11.737562 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d504ddc2-fb6b-4307-9380-3c4fd24ca8db-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d504ddc2-fb6b-4307-9380-3c4fd24ca8db" (UID: "d504ddc2-fb6b-4307-9380-3c4fd24ca8db"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:00:11 crc kubenswrapper[4873]: I1201 09:00:11.766745 4873 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d504ddc2-fb6b-4307-9380-3c4fd24ca8db-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:11 crc kubenswrapper[4873]: I1201 09:00:11.766856 4873 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d504ddc2-fb6b-4307-9380-3c4fd24ca8db-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:11 crc kubenswrapper[4873]: I1201 09:00:11.766870 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z6zkr\" (UniqueName: \"kubernetes.io/projected/d504ddc2-fb6b-4307-9380-3c4fd24ca8db-kube-api-access-z6zkr\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:11 crc kubenswrapper[4873]: I1201 09:00:11.766881 4873 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d504ddc2-fb6b-4307-9380-3c4fd24ca8db-config\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:11 crc kubenswrapper[4873]: I1201 09:00:11.766890 4873 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d504ddc2-fb6b-4307-9380-3c4fd24ca8db-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:12 crc kubenswrapper[4873]: I1201 09:00:12.253677 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d97fcdd8f-qmg5n" event={"ID":"a6a491d2-c38c-41ea-a1bd-ca6f6b276bde","Type":"ContainerStarted","Data":"aa1da172728a3b124cb133ca83e6bf19837b0b45f09faf3ace40ef4e544a21ab"} Dec 01 09:00:12 crc kubenswrapper[4873]: I1201 09:00:12.254114 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d97fcdd8f-qmg5n" Dec 01 09:00:12 crc kubenswrapper[4873]: I1201 09:00:12.265655 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f64d5748f-cjt2l" Dec 01 09:00:12 crc kubenswrapper[4873]: I1201 09:00:12.265644 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f64d5748f-cjt2l" event={"ID":"d504ddc2-fb6b-4307-9380-3c4fd24ca8db","Type":"ContainerDied","Data":"4f608b453c44eed55aa497a873194c0a0b0dd82ecdbec86335dce3b27f4a6182"} Dec 01 09:00:12 crc kubenswrapper[4873]: I1201 09:00:12.265927 4873 scope.go:117] "RemoveContainer" containerID="48e6422ff5a86c5a3fd6771d60481e6825da808c2a512d77ca2c812a65aee862" Dec 01 09:00:12 crc kubenswrapper[4873]: I1201 09:00:12.268741 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b9a7d7f4-a682-4e90-99e4-b3b2afe2c718","Type":"ContainerStarted","Data":"bff8f6c505ec206485d8e43e2255f5b9bafab079664d40472733a9a2f7af25f7"} Dec 01 09:00:12 crc kubenswrapper[4873]: I1201 09:00:12.277927 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-64c5c798d6-h87s6" event={"ID":"c10e1747-9f3e-4fc3-bd55-b83c7a93ea12","Type":"ContainerStarted","Data":"6f76fe202e35950a65f0577471da2e00b98b3506205c60cf2b9d9cb456f3eca0"} Dec 01 09:00:12 crc kubenswrapper[4873]: I1201 09:00:12.279114 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-64c5c798d6-h87s6" Dec 01 09:00:12 crc kubenswrapper[4873]: I1201 09:00:12.279258 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-64c5c798d6-h87s6" Dec 01 09:00:12 crc kubenswrapper[4873]: I1201 09:00:12.285565 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d97fcdd8f-qmg5n" podStartSLOduration=3.285535227 podStartE2EDuration="3.285535227s" podCreationTimestamp="2025-12-01 09:00:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:00:12.27627284 +0000 UTC m=+1188.178381399" watchObservedRunningTime="2025-12-01 09:00:12.285535227 +0000 UTC m=+1188.187643766" Dec 01 09:00:12 crc kubenswrapper[4873]: I1201 09:00:12.319513 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-64c5c798d6-h87s6" podStartSLOduration=3.319487747 podStartE2EDuration="3.319487747s" podCreationTimestamp="2025-12-01 09:00:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:00:12.300604206 +0000 UTC m=+1188.202712755" watchObservedRunningTime="2025-12-01 09:00:12.319487747 +0000 UTC m=+1188.221596276" Dec 01 09:00:12 crc kubenswrapper[4873]: I1201 09:00:12.377375 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f64d5748f-cjt2l"] Dec 01 09:00:12 crc kubenswrapper[4873]: I1201 09:00:12.394631 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f64d5748f-cjt2l"] Dec 01 09:00:12 crc kubenswrapper[4873]: I1201 09:00:12.446692 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d504ddc2-fb6b-4307-9380-3c4fd24ca8db" path="/var/lib/kubelet/pods/d504ddc2-fb6b-4307-9380-3c4fd24ca8db/volumes" Dec 01 09:00:13 crc kubenswrapper[4873]: I1201 09:00:13.308554 4873 generic.go:334] "Generic (PLEG): container finished" podID="e53689a0-5d92-4e62-bb63-3a56af762a59" containerID="0b43d48616acc5de469072536f3377110552ad3499b67b694a963f6c3a326cbc" exitCode=0 Dec 01 09:00:13 crc kubenswrapper[4873]: I1201 09:00:13.309041 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e53689a0-5d92-4e62-bb63-3a56af762a59","Type":"ContainerDied","Data":"0b43d48616acc5de469072536f3377110552ad3499b67b694a963f6c3a326cbc"} Dec 01 09:00:13 crc kubenswrapper[4873]: I1201 09:00:13.325426 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b9a7d7f4-a682-4e90-99e4-b3b2afe2c718","Type":"ContainerStarted","Data":"615aa731dc33f19d05ed890e49230f80996b2913278e90e5c3c38ae4acb42c02"} Dec 01 09:00:13 crc kubenswrapper[4873]: I1201 09:00:13.325997 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="b9a7d7f4-a682-4e90-99e4-b3b2afe2c718" containerName="cinder-api-log" containerID="cri-o://bff8f6c505ec206485d8e43e2255f5b9bafab079664d40472733a9a2f7af25f7" gracePeriod=30 Dec 01 09:00:13 crc kubenswrapper[4873]: I1201 09:00:13.326164 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="b9a7d7f4-a682-4e90-99e4-b3b2afe2c718" containerName="cinder-api" containerID="cri-o://615aa731dc33f19d05ed890e49230f80996b2913278e90e5c3c38ae4acb42c02" gracePeriod=30 Dec 01 09:00:13 crc kubenswrapper[4873]: I1201 09:00:13.326575 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 01 09:00:13 crc kubenswrapper[4873]: I1201 09:00:13.333317 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"11731b74-fe99-40c4-9445-f86102242d05","Type":"ContainerStarted","Data":"e77db932d96e0f5c27d59ad43932b7ce4fd4ec9f33e6f1accbc0146e681dcb29"} Dec 01 09:00:13 crc kubenswrapper[4873]: I1201 09:00:13.358382 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.358347114 podStartE2EDuration="5.358347114s" podCreationTimestamp="2025-12-01 09:00:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:00:13.350532092 +0000 UTC m=+1189.252640631" watchObservedRunningTime="2025-12-01 09:00:13.358347114 +0000 UTC m=+1189.260455653" Dec 01 09:00:13 crc kubenswrapper[4873]: I1201 09:00:13.674804 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 09:00:13 crc kubenswrapper[4873]: I1201 09:00:13.830182 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e53689a0-5d92-4e62-bb63-3a56af762a59-sg-core-conf-yaml\") pod \"e53689a0-5d92-4e62-bb63-3a56af762a59\" (UID: \"e53689a0-5d92-4e62-bb63-3a56af762a59\") " Dec 01 09:00:13 crc kubenswrapper[4873]: I1201 09:00:13.830306 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e53689a0-5d92-4e62-bb63-3a56af762a59-combined-ca-bundle\") pod \"e53689a0-5d92-4e62-bb63-3a56af762a59\" (UID: \"e53689a0-5d92-4e62-bb63-3a56af762a59\") " Dec 01 09:00:13 crc kubenswrapper[4873]: I1201 09:00:13.830356 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e53689a0-5d92-4e62-bb63-3a56af762a59-scripts\") pod \"e53689a0-5d92-4e62-bb63-3a56af762a59\" (UID: \"e53689a0-5d92-4e62-bb63-3a56af762a59\") " Dec 01 09:00:13 crc kubenswrapper[4873]: I1201 09:00:13.830393 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-spv5x\" (UniqueName: \"kubernetes.io/projected/e53689a0-5d92-4e62-bb63-3a56af762a59-kube-api-access-spv5x\") pod \"e53689a0-5d92-4e62-bb63-3a56af762a59\" (UID: \"e53689a0-5d92-4e62-bb63-3a56af762a59\") " Dec 01 09:00:13 crc kubenswrapper[4873]: I1201 09:00:13.830513 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e53689a0-5d92-4e62-bb63-3a56af762a59-log-httpd\") pod \"e53689a0-5d92-4e62-bb63-3a56af762a59\" (UID: \"e53689a0-5d92-4e62-bb63-3a56af762a59\") " Dec 01 09:00:13 crc kubenswrapper[4873]: I1201 09:00:13.830687 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e53689a0-5d92-4e62-bb63-3a56af762a59-config-data\") pod \"e53689a0-5d92-4e62-bb63-3a56af762a59\" (UID: \"e53689a0-5d92-4e62-bb63-3a56af762a59\") " Dec 01 09:00:13 crc kubenswrapper[4873]: I1201 09:00:13.830735 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e53689a0-5d92-4e62-bb63-3a56af762a59-run-httpd\") pod \"e53689a0-5d92-4e62-bb63-3a56af762a59\" (UID: \"e53689a0-5d92-4e62-bb63-3a56af762a59\") " Dec 01 09:00:13 crc kubenswrapper[4873]: I1201 09:00:13.831782 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e53689a0-5d92-4e62-bb63-3a56af762a59-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e53689a0-5d92-4e62-bb63-3a56af762a59" (UID: "e53689a0-5d92-4e62-bb63-3a56af762a59"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:00:13 crc kubenswrapper[4873]: I1201 09:00:13.835154 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e53689a0-5d92-4e62-bb63-3a56af762a59-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e53689a0-5d92-4e62-bb63-3a56af762a59" (UID: "e53689a0-5d92-4e62-bb63-3a56af762a59"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:00:13 crc kubenswrapper[4873]: I1201 09:00:13.841711 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e53689a0-5d92-4e62-bb63-3a56af762a59-scripts" (OuterVolumeSpecName: "scripts") pod "e53689a0-5d92-4e62-bb63-3a56af762a59" (UID: "e53689a0-5d92-4e62-bb63-3a56af762a59"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:13 crc kubenswrapper[4873]: I1201 09:00:13.843588 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e53689a0-5d92-4e62-bb63-3a56af762a59-kube-api-access-spv5x" (OuterVolumeSpecName: "kube-api-access-spv5x") pod "e53689a0-5d92-4e62-bb63-3a56af762a59" (UID: "e53689a0-5d92-4e62-bb63-3a56af762a59"). InnerVolumeSpecName "kube-api-access-spv5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:00:13 crc kubenswrapper[4873]: I1201 09:00:13.878980 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e53689a0-5d92-4e62-bb63-3a56af762a59-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e53689a0-5d92-4e62-bb63-3a56af762a59" (UID: "e53689a0-5d92-4e62-bb63-3a56af762a59"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:13 crc kubenswrapper[4873]: I1201 09:00:13.935304 4873 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e53689a0-5d92-4e62-bb63-3a56af762a59-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:13 crc kubenswrapper[4873]: I1201 09:00:13.935345 4873 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e53689a0-5d92-4e62-bb63-3a56af762a59-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:13 crc kubenswrapper[4873]: I1201 09:00:13.935359 4873 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e53689a0-5d92-4e62-bb63-3a56af762a59-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:13 crc kubenswrapper[4873]: I1201 09:00:13.935369 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-spv5x\" (UniqueName: \"kubernetes.io/projected/e53689a0-5d92-4e62-bb63-3a56af762a59-kube-api-access-spv5x\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:13 crc kubenswrapper[4873]: I1201 09:00:13.935379 4873 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e53689a0-5d92-4e62-bb63-3a56af762a59-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:13 crc kubenswrapper[4873]: I1201 09:00:13.944183 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e53689a0-5d92-4e62-bb63-3a56af762a59-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e53689a0-5d92-4e62-bb63-3a56af762a59" (UID: "e53689a0-5d92-4e62-bb63-3a56af762a59"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:13 crc kubenswrapper[4873]: I1201 09:00:13.972418 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e53689a0-5d92-4e62-bb63-3a56af762a59-config-data" (OuterVolumeSpecName: "config-data") pod "e53689a0-5d92-4e62-bb63-3a56af762a59" (UID: "e53689a0-5d92-4e62-bb63-3a56af762a59"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.044858 4873 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e53689a0-5d92-4e62-bb63-3a56af762a59-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.044918 4873 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e53689a0-5d92-4e62-bb63-3a56af762a59-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.208717 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.247209 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swj9t\" (UniqueName: \"kubernetes.io/projected/b9a7d7f4-a682-4e90-99e4-b3b2afe2c718-kube-api-access-swj9t\") pod \"b9a7d7f4-a682-4e90-99e4-b3b2afe2c718\" (UID: \"b9a7d7f4-a682-4e90-99e4-b3b2afe2c718\") " Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.247283 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9a7d7f4-a682-4e90-99e4-b3b2afe2c718-config-data\") pod \"b9a7d7f4-a682-4e90-99e4-b3b2afe2c718\" (UID: \"b9a7d7f4-a682-4e90-99e4-b3b2afe2c718\") " Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.247363 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b9a7d7f4-a682-4e90-99e4-b3b2afe2c718-logs\") pod \"b9a7d7f4-a682-4e90-99e4-b3b2afe2c718\" (UID: \"b9a7d7f4-a682-4e90-99e4-b3b2afe2c718\") " Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.247388 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9a7d7f4-a682-4e90-99e4-b3b2afe2c718-combined-ca-bundle\") pod \"b9a7d7f4-a682-4e90-99e4-b3b2afe2c718\" (UID: \"b9a7d7f4-a682-4e90-99e4-b3b2afe2c718\") " Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.247428 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b9a7d7f4-a682-4e90-99e4-b3b2afe2c718-etc-machine-id\") pod \"b9a7d7f4-a682-4e90-99e4-b3b2afe2c718\" (UID: \"b9a7d7f4-a682-4e90-99e4-b3b2afe2c718\") " Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.247502 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b9a7d7f4-a682-4e90-99e4-b3b2afe2c718-config-data-custom\") pod \"b9a7d7f4-a682-4e90-99e4-b3b2afe2c718\" (UID: \"b9a7d7f4-a682-4e90-99e4-b3b2afe2c718\") " Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.247547 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9a7d7f4-a682-4e90-99e4-b3b2afe2c718-scripts\") pod \"b9a7d7f4-a682-4e90-99e4-b3b2afe2c718\" (UID: \"b9a7d7f4-a682-4e90-99e4-b3b2afe2c718\") " Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.247850 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b9a7d7f4-a682-4e90-99e4-b3b2afe2c718-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "b9a7d7f4-a682-4e90-99e4-b3b2afe2c718" (UID: "b9a7d7f4-a682-4e90-99e4-b3b2afe2c718"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.248169 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9a7d7f4-a682-4e90-99e4-b3b2afe2c718-logs" (OuterVolumeSpecName: "logs") pod "b9a7d7f4-a682-4e90-99e4-b3b2afe2c718" (UID: "b9a7d7f4-a682-4e90-99e4-b3b2afe2c718"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.262223 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9a7d7f4-a682-4e90-99e4-b3b2afe2c718-kube-api-access-swj9t" (OuterVolumeSpecName: "kube-api-access-swj9t") pod "b9a7d7f4-a682-4e90-99e4-b3b2afe2c718" (UID: "b9a7d7f4-a682-4e90-99e4-b3b2afe2c718"). InnerVolumeSpecName "kube-api-access-swj9t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.268212 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9a7d7f4-a682-4e90-99e4-b3b2afe2c718-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "b9a7d7f4-a682-4e90-99e4-b3b2afe2c718" (UID: "b9a7d7f4-a682-4e90-99e4-b3b2afe2c718"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.272268 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9a7d7f4-a682-4e90-99e4-b3b2afe2c718-scripts" (OuterVolumeSpecName: "scripts") pod "b9a7d7f4-a682-4e90-99e4-b3b2afe2c718" (UID: "b9a7d7f4-a682-4e90-99e4-b3b2afe2c718"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.322680 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9a7d7f4-a682-4e90-99e4-b3b2afe2c718-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b9a7d7f4-a682-4e90-99e4-b3b2afe2c718" (UID: "b9a7d7f4-a682-4e90-99e4-b3b2afe2c718"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.341263 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9a7d7f4-a682-4e90-99e4-b3b2afe2c718-config-data" (OuterVolumeSpecName: "config-data") pod "b9a7d7f4-a682-4e90-99e4-b3b2afe2c718" (UID: "b9a7d7f4-a682-4e90-99e4-b3b2afe2c718"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.348968 4873 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b9a7d7f4-a682-4e90-99e4-b3b2afe2c718-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.348995 4873 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9a7d7f4-a682-4e90-99e4-b3b2afe2c718-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.349007 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swj9t\" (UniqueName: \"kubernetes.io/projected/b9a7d7f4-a682-4e90-99e4-b3b2afe2c718-kube-api-access-swj9t\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.349035 4873 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9a7d7f4-a682-4e90-99e4-b3b2afe2c718-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.349046 4873 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b9a7d7f4-a682-4e90-99e4-b3b2afe2c718-logs\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.349055 4873 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9a7d7f4-a682-4e90-99e4-b3b2afe2c718-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.349065 4873 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b9a7d7f4-a682-4e90-99e4-b3b2afe2c718-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.383496 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-9b48ddc7c-fn75z" event={"ID":"4d0787b5-0da8-493b-b55d-b6e36a759f7f","Type":"ContainerStarted","Data":"3d7d239994212f69b19cc2cad35629147dd05a82a6f32cf7238a0885bf2065e8"} Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.383567 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-9b48ddc7c-fn75z" event={"ID":"4d0787b5-0da8-493b-b55d-b6e36a759f7f","Type":"ContainerStarted","Data":"85913f762ef7a883d7aa8220163d99ddaf9e109734f38c86eff8c55e70b1f10b"} Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.388915 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e53689a0-5d92-4e62-bb63-3a56af762a59","Type":"ContainerDied","Data":"98ad9d5515c70fcee8f018fb5f0cfdb6c0e0b5d6fcc8e7d6ed9338d994e68774"} Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.388962 4873 scope.go:117] "RemoveContainer" containerID="068b339c05967459adb9767502db98e3ed7f8097f9d5a942acefdc2ee1b56026" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.388973 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.398355 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6674fb675b-58jdf" event={"ID":"99f7daba-8483-4494-b783-9628f828ea49","Type":"ContainerStarted","Data":"8df8021d54b794b7b830a03291c379310ecd54b9c6af455fc073d29673696042"} Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.398401 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6674fb675b-58jdf" event={"ID":"99f7daba-8483-4494-b783-9628f828ea49","Type":"ContainerStarted","Data":"9df8b40a24879dd009d143073d9e3e414295ed3d684c6a438850de5561f00cd7"} Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.413858 4873 generic.go:334] "Generic (PLEG): container finished" podID="b9a7d7f4-a682-4e90-99e4-b3b2afe2c718" containerID="615aa731dc33f19d05ed890e49230f80996b2913278e90e5c3c38ae4acb42c02" exitCode=0 Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.413901 4873 generic.go:334] "Generic (PLEG): container finished" podID="b9a7d7f4-a682-4e90-99e4-b3b2afe2c718" containerID="bff8f6c505ec206485d8e43e2255f5b9bafab079664d40472733a9a2f7af25f7" exitCode=143 Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.414057 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.414440 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b9a7d7f4-a682-4e90-99e4-b3b2afe2c718","Type":"ContainerDied","Data":"615aa731dc33f19d05ed890e49230f80996b2913278e90e5c3c38ae4acb42c02"} Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.414517 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b9a7d7f4-a682-4e90-99e4-b3b2afe2c718","Type":"ContainerDied","Data":"bff8f6c505ec206485d8e43e2255f5b9bafab079664d40472733a9a2f7af25f7"} Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.414534 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b9a7d7f4-a682-4e90-99e4-b3b2afe2c718","Type":"ContainerDied","Data":"818f3fa6d352a247655200dfe50d9ebb7fb5071a15a3cfcb04b3cddad6efedbe"} Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.422205 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"11731b74-fe99-40c4-9445-f86102242d05","Type":"ContainerStarted","Data":"c4008ccedc96a77868ded5c3f956a834a5990ecf0d660b15d7bb6eb8270f46e7"} Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.422361 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-9b48ddc7c-fn75z" podStartSLOduration=2.958697722 podStartE2EDuration="5.422332113s" podCreationTimestamp="2025-12-01 09:00:09 +0000 UTC" firstStartedPulling="2025-12-01 09:00:10.535244842 +0000 UTC m=+1186.437353381" lastFinishedPulling="2025-12-01 09:00:12.998879233 +0000 UTC m=+1188.900987772" observedRunningTime="2025-12-01 09:00:14.416589393 +0000 UTC m=+1190.318697932" watchObservedRunningTime="2025-12-01 09:00:14.422332113 +0000 UTC m=+1190.324440652" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.433248 4873 scope.go:117] "RemoveContainer" containerID="b33b0f231101926f73eefa68214880e4c14df8323cb5ce1c1750a3ab4b4e019d" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.479618 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-6674fb675b-58jdf" podStartSLOduration=3.027871545 podStartE2EDuration="5.479584253s" podCreationTimestamp="2025-12-01 09:00:09 +0000 UTC" firstStartedPulling="2025-12-01 09:00:10.534767761 +0000 UTC m=+1186.436876300" lastFinishedPulling="2025-12-01 09:00:12.986480469 +0000 UTC m=+1188.888589008" observedRunningTime="2025-12-01 09:00:14.455803202 +0000 UTC m=+1190.357911741" watchObservedRunningTime="2025-12-01 09:00:14.479584253 +0000 UTC m=+1190.381692792" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.489931 4873 scope.go:117] "RemoveContainer" containerID="0b43d48616acc5de469072536f3377110552ad3499b67b694a963f6c3a326cbc" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.500545 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.518259 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.531757 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=5.128612735 podStartE2EDuration="6.531732349s" podCreationTimestamp="2025-12-01 09:00:08 +0000 UTC" firstStartedPulling="2025-12-01 09:00:10.096819211 +0000 UTC m=+1185.998927750" lastFinishedPulling="2025-12-01 09:00:11.499938815 +0000 UTC m=+1187.402047364" observedRunningTime="2025-12-01 09:00:14.515094482 +0000 UTC m=+1190.417203021" watchObservedRunningTime="2025-12-01 09:00:14.531732349 +0000 UTC m=+1190.433840888" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.552344 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 09:00:14 crc kubenswrapper[4873]: E1201 09:00:14.552760 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e53689a0-5d92-4e62-bb63-3a56af762a59" containerName="ceilometer-notification-agent" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.552777 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="e53689a0-5d92-4e62-bb63-3a56af762a59" containerName="ceilometer-notification-agent" Dec 01 09:00:14 crc kubenswrapper[4873]: E1201 09:00:14.552799 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9a7d7f4-a682-4e90-99e4-b3b2afe2c718" containerName="cinder-api" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.552805 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9a7d7f4-a682-4e90-99e4-b3b2afe2c718" containerName="cinder-api" Dec 01 09:00:14 crc kubenswrapper[4873]: E1201 09:00:14.552814 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d504ddc2-fb6b-4307-9380-3c4fd24ca8db" containerName="init" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.552821 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="d504ddc2-fb6b-4307-9380-3c4fd24ca8db" containerName="init" Dec 01 09:00:14 crc kubenswrapper[4873]: E1201 09:00:14.552829 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9a7d7f4-a682-4e90-99e4-b3b2afe2c718" containerName="cinder-api-log" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.552835 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9a7d7f4-a682-4e90-99e4-b3b2afe2c718" containerName="cinder-api-log" Dec 01 09:00:14 crc kubenswrapper[4873]: E1201 09:00:14.552847 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e53689a0-5d92-4e62-bb63-3a56af762a59" containerName="sg-core" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.552852 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="e53689a0-5d92-4e62-bb63-3a56af762a59" containerName="sg-core" Dec 01 09:00:14 crc kubenswrapper[4873]: E1201 09:00:14.552864 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e53689a0-5d92-4e62-bb63-3a56af762a59" containerName="ceilometer-central-agent" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.552873 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="e53689a0-5d92-4e62-bb63-3a56af762a59" containerName="ceilometer-central-agent" Dec 01 09:00:14 crc kubenswrapper[4873]: E1201 09:00:14.552889 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e53689a0-5d92-4e62-bb63-3a56af762a59" containerName="proxy-httpd" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.552895 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="e53689a0-5d92-4e62-bb63-3a56af762a59" containerName="proxy-httpd" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.553067 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9a7d7f4-a682-4e90-99e4-b3b2afe2c718" containerName="cinder-api-log" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.553085 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="e53689a0-5d92-4e62-bb63-3a56af762a59" containerName="sg-core" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.553094 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="e53689a0-5d92-4e62-bb63-3a56af762a59" containerName="ceilometer-central-agent" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.553106 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="d504ddc2-fb6b-4307-9380-3c4fd24ca8db" containerName="init" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.553119 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="e53689a0-5d92-4e62-bb63-3a56af762a59" containerName="ceilometer-notification-agent" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.553128 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9a7d7f4-a682-4e90-99e4-b3b2afe2c718" containerName="cinder-api" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.553137 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="e53689a0-5d92-4e62-bb63-3a56af762a59" containerName="proxy-httpd" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.556915 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.559469 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.559689 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.563030 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/15c2d570-e3c5-42e4-b837-b188b167be1f-log-httpd\") pod \"ceilometer-0\" (UID: \"15c2d570-e3c5-42e4-b837-b188b167be1f\") " pod="openstack/ceilometer-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.563091 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/15c2d570-e3c5-42e4-b837-b188b167be1f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"15c2d570-e3c5-42e4-b837-b188b167be1f\") " pod="openstack/ceilometer-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.563116 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15c2d570-e3c5-42e4-b837-b188b167be1f-config-data\") pod \"ceilometer-0\" (UID: \"15c2d570-e3c5-42e4-b837-b188b167be1f\") " pod="openstack/ceilometer-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.563138 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8ct2\" (UniqueName: \"kubernetes.io/projected/15c2d570-e3c5-42e4-b837-b188b167be1f-kube-api-access-p8ct2\") pod \"ceilometer-0\" (UID: \"15c2d570-e3c5-42e4-b837-b188b167be1f\") " pod="openstack/ceilometer-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.563174 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15c2d570-e3c5-42e4-b837-b188b167be1f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"15c2d570-e3c5-42e4-b837-b188b167be1f\") " pod="openstack/ceilometer-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.563210 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/15c2d570-e3c5-42e4-b837-b188b167be1f-run-httpd\") pod \"ceilometer-0\" (UID: \"15c2d570-e3c5-42e4-b837-b188b167be1f\") " pod="openstack/ceilometer-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.563256 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15c2d570-e3c5-42e4-b837-b188b167be1f-scripts\") pod \"ceilometer-0\" (UID: \"15c2d570-e3c5-42e4-b837-b188b167be1f\") " pod="openstack/ceilometer-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.570346 4873 scope.go:117] "RemoveContainer" containerID="31940bfd155d788c99f80b4389ebba5de7bb5078293bf263879bda05942db53b" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.572163 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.598562 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.605480 4873 scope.go:117] "RemoveContainer" containerID="615aa731dc33f19d05ed890e49230f80996b2913278e90e5c3c38ae4acb42c02" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.627969 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.633148 4873 scope.go:117] "RemoveContainer" containerID="bff8f6c505ec206485d8e43e2255f5b9bafab079664d40472733a9a2f7af25f7" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.648295 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.649917 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.653399 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.653636 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.653840 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.661093 4873 scope.go:117] "RemoveContainer" containerID="615aa731dc33f19d05ed890e49230f80996b2913278e90e5c3c38ae4acb42c02" Dec 01 09:00:14 crc kubenswrapper[4873]: E1201 09:00:14.662079 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"615aa731dc33f19d05ed890e49230f80996b2913278e90e5c3c38ae4acb42c02\": container with ID starting with 615aa731dc33f19d05ed890e49230f80996b2913278e90e5c3c38ae4acb42c02 not found: ID does not exist" containerID="615aa731dc33f19d05ed890e49230f80996b2913278e90e5c3c38ae4acb42c02" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.662145 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"615aa731dc33f19d05ed890e49230f80996b2913278e90e5c3c38ae4acb42c02"} err="failed to get container status \"615aa731dc33f19d05ed890e49230f80996b2913278e90e5c3c38ae4acb42c02\": rpc error: code = NotFound desc = could not find container \"615aa731dc33f19d05ed890e49230f80996b2913278e90e5c3c38ae4acb42c02\": container with ID starting with 615aa731dc33f19d05ed890e49230f80996b2913278e90e5c3c38ae4acb42c02 not found: ID does not exist" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.662206 4873 scope.go:117] "RemoveContainer" containerID="bff8f6c505ec206485d8e43e2255f5b9bafab079664d40472733a9a2f7af25f7" Dec 01 09:00:14 crc kubenswrapper[4873]: E1201 09:00:14.663572 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bff8f6c505ec206485d8e43e2255f5b9bafab079664d40472733a9a2f7af25f7\": container with ID starting with bff8f6c505ec206485d8e43e2255f5b9bafab079664d40472733a9a2f7af25f7 not found: ID does not exist" containerID="bff8f6c505ec206485d8e43e2255f5b9bafab079664d40472733a9a2f7af25f7" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.663629 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bff8f6c505ec206485d8e43e2255f5b9bafab079664d40472733a9a2f7af25f7"} err="failed to get container status \"bff8f6c505ec206485d8e43e2255f5b9bafab079664d40472733a9a2f7af25f7\": rpc error: code = NotFound desc = could not find container \"bff8f6c505ec206485d8e43e2255f5b9bafab079664d40472733a9a2f7af25f7\": container with ID starting with bff8f6c505ec206485d8e43e2255f5b9bafab079664d40472733a9a2f7af25f7 not found: ID does not exist" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.663664 4873 scope.go:117] "RemoveContainer" containerID="615aa731dc33f19d05ed890e49230f80996b2913278e90e5c3c38ae4acb42c02" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.664374 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"615aa731dc33f19d05ed890e49230f80996b2913278e90e5c3c38ae4acb42c02"} err="failed to get container status \"615aa731dc33f19d05ed890e49230f80996b2913278e90e5c3c38ae4acb42c02\": rpc error: code = NotFound desc = could not find container \"615aa731dc33f19d05ed890e49230f80996b2913278e90e5c3c38ae4acb42c02\": container with ID starting with 615aa731dc33f19d05ed890e49230f80996b2913278e90e5c3c38ae4acb42c02 not found: ID does not exist" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.664419 4873 scope.go:117] "RemoveContainer" containerID="bff8f6c505ec206485d8e43e2255f5b9bafab079664d40472733a9a2f7af25f7" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.665819 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bff8f6c505ec206485d8e43e2255f5b9bafab079664d40472733a9a2f7af25f7"} err="failed to get container status \"bff8f6c505ec206485d8e43e2255f5b9bafab079664d40472733a9a2f7af25f7\": rpc error: code = NotFound desc = could not find container \"bff8f6c505ec206485d8e43e2255f5b9bafab079664d40472733a9a2f7af25f7\": container with ID starting with bff8f6c505ec206485d8e43e2255f5b9bafab079664d40472733a9a2f7af25f7 not found: ID does not exist" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.667825 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/15c2d570-e3c5-42e4-b837-b188b167be1f-log-httpd\") pod \"ceilometer-0\" (UID: \"15c2d570-e3c5-42e4-b837-b188b167be1f\") " pod="openstack/ceilometer-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.667876 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hds7w\" (UniqueName: \"kubernetes.io/projected/50bc343c-5aca-46d3-a9de-31546ac7c45f-kube-api-access-hds7w\") pod \"cinder-api-0\" (UID: \"50bc343c-5aca-46d3-a9de-31546ac7c45f\") " pod="openstack/cinder-api-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.667932 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/50bc343c-5aca-46d3-a9de-31546ac7c45f-config-data-custom\") pod \"cinder-api-0\" (UID: \"50bc343c-5aca-46d3-a9de-31546ac7c45f\") " pod="openstack/cinder-api-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.667953 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/50bc343c-5aca-46d3-a9de-31546ac7c45f-public-tls-certs\") pod \"cinder-api-0\" (UID: \"50bc343c-5aca-46d3-a9de-31546ac7c45f\") " pod="openstack/cinder-api-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.667981 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/15c2d570-e3c5-42e4-b837-b188b167be1f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"15c2d570-e3c5-42e4-b837-b188b167be1f\") " pod="openstack/ceilometer-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.667998 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/50bc343c-5aca-46d3-a9de-31546ac7c45f-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"50bc343c-5aca-46d3-a9de-31546ac7c45f\") " pod="openstack/cinder-api-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.668097 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15c2d570-e3c5-42e4-b837-b188b167be1f-config-data\") pod \"ceilometer-0\" (UID: \"15c2d570-e3c5-42e4-b837-b188b167be1f\") " pod="openstack/ceilometer-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.668119 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8ct2\" (UniqueName: \"kubernetes.io/projected/15c2d570-e3c5-42e4-b837-b188b167be1f-kube-api-access-p8ct2\") pod \"ceilometer-0\" (UID: \"15c2d570-e3c5-42e4-b837-b188b167be1f\") " pod="openstack/ceilometer-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.668154 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/50bc343c-5aca-46d3-a9de-31546ac7c45f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"50bc343c-5aca-46d3-a9de-31546ac7c45f\") " pod="openstack/cinder-api-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.668194 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50bc343c-5aca-46d3-a9de-31546ac7c45f-config-data\") pod \"cinder-api-0\" (UID: \"50bc343c-5aca-46d3-a9de-31546ac7c45f\") " pod="openstack/cinder-api-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.668219 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15c2d570-e3c5-42e4-b837-b188b167be1f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"15c2d570-e3c5-42e4-b837-b188b167be1f\") " pod="openstack/ceilometer-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.668271 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/15c2d570-e3c5-42e4-b837-b188b167be1f-run-httpd\") pod \"ceilometer-0\" (UID: \"15c2d570-e3c5-42e4-b837-b188b167be1f\") " pod="openstack/ceilometer-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.668301 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50bc343c-5aca-46d3-a9de-31546ac7c45f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"50bc343c-5aca-46d3-a9de-31546ac7c45f\") " pod="openstack/cinder-api-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.668332 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/15c2d570-e3c5-42e4-b837-b188b167be1f-log-httpd\") pod \"ceilometer-0\" (UID: \"15c2d570-e3c5-42e4-b837-b188b167be1f\") " pod="openstack/ceilometer-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.668361 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15c2d570-e3c5-42e4-b837-b188b167be1f-scripts\") pod \"ceilometer-0\" (UID: \"15c2d570-e3c5-42e4-b837-b188b167be1f\") " pod="openstack/ceilometer-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.668405 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50bc343c-5aca-46d3-a9de-31546ac7c45f-logs\") pod \"cinder-api-0\" (UID: \"50bc343c-5aca-46d3-a9de-31546ac7c45f\") " pod="openstack/cinder-api-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.668438 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50bc343c-5aca-46d3-a9de-31546ac7c45f-scripts\") pod \"cinder-api-0\" (UID: \"50bc343c-5aca-46d3-a9de-31546ac7c45f\") " pod="openstack/cinder-api-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.669653 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/15c2d570-e3c5-42e4-b837-b188b167be1f-run-httpd\") pod \"ceilometer-0\" (UID: \"15c2d570-e3c5-42e4-b837-b188b167be1f\") " pod="openstack/ceilometer-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.669661 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.675501 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15c2d570-e3c5-42e4-b837-b188b167be1f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"15c2d570-e3c5-42e4-b837-b188b167be1f\") " pod="openstack/ceilometer-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.681980 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15c2d570-e3c5-42e4-b837-b188b167be1f-scripts\") pod \"ceilometer-0\" (UID: \"15c2d570-e3c5-42e4-b837-b188b167be1f\") " pod="openstack/ceilometer-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.682723 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/15c2d570-e3c5-42e4-b837-b188b167be1f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"15c2d570-e3c5-42e4-b837-b188b167be1f\") " pod="openstack/ceilometer-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.683170 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15c2d570-e3c5-42e4-b837-b188b167be1f-config-data\") pod \"ceilometer-0\" (UID: \"15c2d570-e3c5-42e4-b837-b188b167be1f\") " pod="openstack/ceilometer-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.701941 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8ct2\" (UniqueName: \"kubernetes.io/projected/15c2d570-e3c5-42e4-b837-b188b167be1f-kube-api-access-p8ct2\") pod \"ceilometer-0\" (UID: \"15c2d570-e3c5-42e4-b837-b188b167be1f\") " pod="openstack/ceilometer-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.769861 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50bc343c-5aca-46d3-a9de-31546ac7c45f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"50bc343c-5aca-46d3-a9de-31546ac7c45f\") " pod="openstack/cinder-api-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.769978 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50bc343c-5aca-46d3-a9de-31546ac7c45f-logs\") pod \"cinder-api-0\" (UID: \"50bc343c-5aca-46d3-a9de-31546ac7c45f\") " pod="openstack/cinder-api-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.770015 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50bc343c-5aca-46d3-a9de-31546ac7c45f-scripts\") pod \"cinder-api-0\" (UID: \"50bc343c-5aca-46d3-a9de-31546ac7c45f\") " pod="openstack/cinder-api-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.770075 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hds7w\" (UniqueName: \"kubernetes.io/projected/50bc343c-5aca-46d3-a9de-31546ac7c45f-kube-api-access-hds7w\") pod \"cinder-api-0\" (UID: \"50bc343c-5aca-46d3-a9de-31546ac7c45f\") " pod="openstack/cinder-api-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.770111 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/50bc343c-5aca-46d3-a9de-31546ac7c45f-config-data-custom\") pod \"cinder-api-0\" (UID: \"50bc343c-5aca-46d3-a9de-31546ac7c45f\") " pod="openstack/cinder-api-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.770126 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/50bc343c-5aca-46d3-a9de-31546ac7c45f-public-tls-certs\") pod \"cinder-api-0\" (UID: \"50bc343c-5aca-46d3-a9de-31546ac7c45f\") " pod="openstack/cinder-api-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.770145 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/50bc343c-5aca-46d3-a9de-31546ac7c45f-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"50bc343c-5aca-46d3-a9de-31546ac7c45f\") " pod="openstack/cinder-api-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.770168 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/50bc343c-5aca-46d3-a9de-31546ac7c45f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"50bc343c-5aca-46d3-a9de-31546ac7c45f\") " pod="openstack/cinder-api-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.770191 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50bc343c-5aca-46d3-a9de-31546ac7c45f-config-data\") pod \"cinder-api-0\" (UID: \"50bc343c-5aca-46d3-a9de-31546ac7c45f\") " pod="openstack/cinder-api-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.771367 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/50bc343c-5aca-46d3-a9de-31546ac7c45f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"50bc343c-5aca-46d3-a9de-31546ac7c45f\") " pod="openstack/cinder-api-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.771636 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50bc343c-5aca-46d3-a9de-31546ac7c45f-logs\") pod \"cinder-api-0\" (UID: \"50bc343c-5aca-46d3-a9de-31546ac7c45f\") " pod="openstack/cinder-api-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.775630 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/50bc343c-5aca-46d3-a9de-31546ac7c45f-config-data-custom\") pod \"cinder-api-0\" (UID: \"50bc343c-5aca-46d3-a9de-31546ac7c45f\") " pod="openstack/cinder-api-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.782675 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50bc343c-5aca-46d3-a9de-31546ac7c45f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"50bc343c-5aca-46d3-a9de-31546ac7c45f\") " pod="openstack/cinder-api-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.788842 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/50bc343c-5aca-46d3-a9de-31546ac7c45f-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"50bc343c-5aca-46d3-a9de-31546ac7c45f\") " pod="openstack/cinder-api-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.788957 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/50bc343c-5aca-46d3-a9de-31546ac7c45f-public-tls-certs\") pod \"cinder-api-0\" (UID: \"50bc343c-5aca-46d3-a9de-31546ac7c45f\") " pod="openstack/cinder-api-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.789302 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50bc343c-5aca-46d3-a9de-31546ac7c45f-scripts\") pod \"cinder-api-0\" (UID: \"50bc343c-5aca-46d3-a9de-31546ac7c45f\") " pod="openstack/cinder-api-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.789325 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50bc343c-5aca-46d3-a9de-31546ac7c45f-config-data\") pod \"cinder-api-0\" (UID: \"50bc343c-5aca-46d3-a9de-31546ac7c45f\") " pod="openstack/cinder-api-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.793089 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hds7w\" (UniqueName: \"kubernetes.io/projected/50bc343c-5aca-46d3-a9de-31546ac7c45f-kube-api-access-hds7w\") pod \"cinder-api-0\" (UID: \"50bc343c-5aca-46d3-a9de-31546ac7c45f\") " pod="openstack/cinder-api-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.912444 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 09:00:14 crc kubenswrapper[4873]: I1201 09:00:14.973074 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 09:00:15 crc kubenswrapper[4873]: I1201 09:00:15.377188 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7fc6f67df6-72xpt"] Dec 01 09:00:15 crc kubenswrapper[4873]: I1201 09:00:15.379779 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7fc6f67df6-72xpt" Dec 01 09:00:15 crc kubenswrapper[4873]: I1201 09:00:15.392386 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 01 09:00:15 crc kubenswrapper[4873]: I1201 09:00:15.392563 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 01 09:00:15 crc kubenswrapper[4873]: I1201 09:00:15.400370 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7fc6f67df6-72xpt"] Dec 01 09:00:15 crc kubenswrapper[4873]: I1201 09:00:15.492096 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/145a11d3-f1db-45aa-a3fe-0d4709905406-config-data-custom\") pod \"barbican-api-7fc6f67df6-72xpt\" (UID: \"145a11d3-f1db-45aa-a3fe-0d4709905406\") " pod="openstack/barbican-api-7fc6f67df6-72xpt" Dec 01 09:00:15 crc kubenswrapper[4873]: I1201 09:00:15.492145 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/145a11d3-f1db-45aa-a3fe-0d4709905406-combined-ca-bundle\") pod \"barbican-api-7fc6f67df6-72xpt\" (UID: \"145a11d3-f1db-45aa-a3fe-0d4709905406\") " pod="openstack/barbican-api-7fc6f67df6-72xpt" Dec 01 09:00:15 crc kubenswrapper[4873]: I1201 09:00:15.492205 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/145a11d3-f1db-45aa-a3fe-0d4709905406-public-tls-certs\") pod \"barbican-api-7fc6f67df6-72xpt\" (UID: \"145a11d3-f1db-45aa-a3fe-0d4709905406\") " pod="openstack/barbican-api-7fc6f67df6-72xpt" Dec 01 09:00:15 crc kubenswrapper[4873]: I1201 09:00:15.492257 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/145a11d3-f1db-45aa-a3fe-0d4709905406-internal-tls-certs\") pod \"barbican-api-7fc6f67df6-72xpt\" (UID: \"145a11d3-f1db-45aa-a3fe-0d4709905406\") " pod="openstack/barbican-api-7fc6f67df6-72xpt" Dec 01 09:00:15 crc kubenswrapper[4873]: I1201 09:00:15.492290 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/145a11d3-f1db-45aa-a3fe-0d4709905406-logs\") pod \"barbican-api-7fc6f67df6-72xpt\" (UID: \"145a11d3-f1db-45aa-a3fe-0d4709905406\") " pod="openstack/barbican-api-7fc6f67df6-72xpt" Dec 01 09:00:15 crc kubenswrapper[4873]: I1201 09:00:15.492341 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/145a11d3-f1db-45aa-a3fe-0d4709905406-config-data\") pod \"barbican-api-7fc6f67df6-72xpt\" (UID: \"145a11d3-f1db-45aa-a3fe-0d4709905406\") " pod="openstack/barbican-api-7fc6f67df6-72xpt" Dec 01 09:00:15 crc kubenswrapper[4873]: I1201 09:00:15.492372 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbwff\" (UniqueName: \"kubernetes.io/projected/145a11d3-f1db-45aa-a3fe-0d4709905406-kube-api-access-lbwff\") pod \"barbican-api-7fc6f67df6-72xpt\" (UID: \"145a11d3-f1db-45aa-a3fe-0d4709905406\") " pod="openstack/barbican-api-7fc6f67df6-72xpt" Dec 01 09:00:15 crc kubenswrapper[4873]: I1201 09:00:15.497886 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 09:00:15 crc kubenswrapper[4873]: I1201 09:00:15.606903 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/145a11d3-f1db-45aa-a3fe-0d4709905406-config-data\") pod \"barbican-api-7fc6f67df6-72xpt\" (UID: \"145a11d3-f1db-45aa-a3fe-0d4709905406\") " pod="openstack/barbican-api-7fc6f67df6-72xpt" Dec 01 09:00:15 crc kubenswrapper[4873]: I1201 09:00:15.606989 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbwff\" (UniqueName: \"kubernetes.io/projected/145a11d3-f1db-45aa-a3fe-0d4709905406-kube-api-access-lbwff\") pod \"barbican-api-7fc6f67df6-72xpt\" (UID: \"145a11d3-f1db-45aa-a3fe-0d4709905406\") " pod="openstack/barbican-api-7fc6f67df6-72xpt" Dec 01 09:00:15 crc kubenswrapper[4873]: I1201 09:00:15.607051 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/145a11d3-f1db-45aa-a3fe-0d4709905406-config-data-custom\") pod \"barbican-api-7fc6f67df6-72xpt\" (UID: \"145a11d3-f1db-45aa-a3fe-0d4709905406\") " pod="openstack/barbican-api-7fc6f67df6-72xpt" Dec 01 09:00:15 crc kubenswrapper[4873]: I1201 09:00:15.607093 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/145a11d3-f1db-45aa-a3fe-0d4709905406-combined-ca-bundle\") pod \"barbican-api-7fc6f67df6-72xpt\" (UID: \"145a11d3-f1db-45aa-a3fe-0d4709905406\") " pod="openstack/barbican-api-7fc6f67df6-72xpt" Dec 01 09:00:15 crc kubenswrapper[4873]: I1201 09:00:15.607152 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/145a11d3-f1db-45aa-a3fe-0d4709905406-public-tls-certs\") pod \"barbican-api-7fc6f67df6-72xpt\" (UID: \"145a11d3-f1db-45aa-a3fe-0d4709905406\") " pod="openstack/barbican-api-7fc6f67df6-72xpt" Dec 01 09:00:15 crc kubenswrapper[4873]: I1201 09:00:15.607327 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/145a11d3-f1db-45aa-a3fe-0d4709905406-internal-tls-certs\") pod \"barbican-api-7fc6f67df6-72xpt\" (UID: \"145a11d3-f1db-45aa-a3fe-0d4709905406\") " pod="openstack/barbican-api-7fc6f67df6-72xpt" Dec 01 09:00:15 crc kubenswrapper[4873]: I1201 09:00:15.607383 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/145a11d3-f1db-45aa-a3fe-0d4709905406-logs\") pod \"barbican-api-7fc6f67df6-72xpt\" (UID: \"145a11d3-f1db-45aa-a3fe-0d4709905406\") " pod="openstack/barbican-api-7fc6f67df6-72xpt" Dec 01 09:00:15 crc kubenswrapper[4873]: I1201 09:00:15.608615 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/145a11d3-f1db-45aa-a3fe-0d4709905406-logs\") pod \"barbican-api-7fc6f67df6-72xpt\" (UID: \"145a11d3-f1db-45aa-a3fe-0d4709905406\") " pod="openstack/barbican-api-7fc6f67df6-72xpt" Dec 01 09:00:15 crc kubenswrapper[4873]: I1201 09:00:15.624362 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/145a11d3-f1db-45aa-a3fe-0d4709905406-public-tls-certs\") pod \"barbican-api-7fc6f67df6-72xpt\" (UID: \"145a11d3-f1db-45aa-a3fe-0d4709905406\") " pod="openstack/barbican-api-7fc6f67df6-72xpt" Dec 01 09:00:15 crc kubenswrapper[4873]: I1201 09:00:15.625334 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/145a11d3-f1db-45aa-a3fe-0d4709905406-internal-tls-certs\") pod \"barbican-api-7fc6f67df6-72xpt\" (UID: \"145a11d3-f1db-45aa-a3fe-0d4709905406\") " pod="openstack/barbican-api-7fc6f67df6-72xpt" Dec 01 09:00:15 crc kubenswrapper[4873]: I1201 09:00:15.626402 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/145a11d3-f1db-45aa-a3fe-0d4709905406-config-data\") pod \"barbican-api-7fc6f67df6-72xpt\" (UID: \"145a11d3-f1db-45aa-a3fe-0d4709905406\") " pod="openstack/barbican-api-7fc6f67df6-72xpt" Dec 01 09:00:15 crc kubenswrapper[4873]: I1201 09:00:15.627722 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/145a11d3-f1db-45aa-a3fe-0d4709905406-combined-ca-bundle\") pod \"barbican-api-7fc6f67df6-72xpt\" (UID: \"145a11d3-f1db-45aa-a3fe-0d4709905406\") " pod="openstack/barbican-api-7fc6f67df6-72xpt" Dec 01 09:00:15 crc kubenswrapper[4873]: I1201 09:00:15.629530 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/145a11d3-f1db-45aa-a3fe-0d4709905406-config-data-custom\") pod \"barbican-api-7fc6f67df6-72xpt\" (UID: \"145a11d3-f1db-45aa-a3fe-0d4709905406\") " pod="openstack/barbican-api-7fc6f67df6-72xpt" Dec 01 09:00:15 crc kubenswrapper[4873]: I1201 09:00:15.650893 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbwff\" (UniqueName: \"kubernetes.io/projected/145a11d3-f1db-45aa-a3fe-0d4709905406-kube-api-access-lbwff\") pod \"barbican-api-7fc6f67df6-72xpt\" (UID: \"145a11d3-f1db-45aa-a3fe-0d4709905406\") " pod="openstack/barbican-api-7fc6f67df6-72xpt" Dec 01 09:00:15 crc kubenswrapper[4873]: I1201 09:00:15.696941 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 01 09:00:15 crc kubenswrapper[4873]: I1201 09:00:15.727763 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7fc6f67df6-72xpt" Dec 01 09:00:16 crc kubenswrapper[4873]: I1201 09:00:16.446364 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9a7d7f4-a682-4e90-99e4-b3b2afe2c718" path="/var/lib/kubelet/pods/b9a7d7f4-a682-4e90-99e4-b3b2afe2c718/volumes" Dec 01 09:00:16 crc kubenswrapper[4873]: I1201 09:00:16.448147 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e53689a0-5d92-4e62-bb63-3a56af762a59" path="/var/lib/kubelet/pods/e53689a0-5d92-4e62-bb63-3a56af762a59/volumes" Dec 01 09:00:16 crc kubenswrapper[4873]: I1201 09:00:16.453743 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7fc6f67df6-72xpt"] Dec 01 09:00:16 crc kubenswrapper[4873]: I1201 09:00:16.482703 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"15c2d570-e3c5-42e4-b837-b188b167be1f","Type":"ContainerStarted","Data":"f2209bbb291eace67e6c2a39ebb72b7ba687e697d507435485f29499349fecb0"} Dec 01 09:00:16 crc kubenswrapper[4873]: I1201 09:00:16.485506 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7fc6f67df6-72xpt" event={"ID":"145a11d3-f1db-45aa-a3fe-0d4709905406","Type":"ContainerStarted","Data":"33e2c355af3de1846a9fac5a92838c57f9d718a59f76334cf799dcbcf9d38cd7"} Dec 01 09:00:16 crc kubenswrapper[4873]: I1201 09:00:16.488224 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"50bc343c-5aca-46d3-a9de-31546ac7c45f","Type":"ContainerStarted","Data":"0f3cea6c69cf24dc21765b2dfdd66ba435ebd8322ff45a709e8d7f0ea035d1a8"} Dec 01 09:00:17 crc kubenswrapper[4873]: I1201 09:00:17.498262 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7fc6f67df6-72xpt" event={"ID":"145a11d3-f1db-45aa-a3fe-0d4709905406","Type":"ContainerStarted","Data":"a12c7f3fae344e44fcee6927e51475b47b20703952f30a1f58a2b72c89eed731"} Dec 01 09:00:17 crc kubenswrapper[4873]: I1201 09:00:17.500270 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7fc6f67df6-72xpt" event={"ID":"145a11d3-f1db-45aa-a3fe-0d4709905406","Type":"ContainerStarted","Data":"3bb8f42766e63d52155f582c5320bf9b2c9975152e7c2661c496c7b8a0ae6c8f"} Dec 01 09:00:17 crc kubenswrapper[4873]: I1201 09:00:17.500326 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"50bc343c-5aca-46d3-a9de-31546ac7c45f","Type":"ContainerStarted","Data":"a353b08de2e817f7c6a8b5121f9f49c51e965846f4fc9e4ce2c3f2b2503f469e"} Dec 01 09:00:17 crc kubenswrapper[4873]: I1201 09:00:17.500345 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"50bc343c-5aca-46d3-a9de-31546ac7c45f","Type":"ContainerStarted","Data":"05f6cf12989201a34a871f46e4b98d72513b26231a25fcfb1f909983834611de"} Dec 01 09:00:17 crc kubenswrapper[4873]: I1201 09:00:17.500406 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7fc6f67df6-72xpt" Dec 01 09:00:17 crc kubenswrapper[4873]: I1201 09:00:17.500423 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7fc6f67df6-72xpt" Dec 01 09:00:17 crc kubenswrapper[4873]: I1201 09:00:17.500438 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 01 09:00:17 crc kubenswrapper[4873]: I1201 09:00:17.503094 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"15c2d570-e3c5-42e4-b837-b188b167be1f","Type":"ContainerStarted","Data":"f21f06a3b000822f1f73e94cdc69477bc1636249d7d3ec510cf2b5a38321bc7b"} Dec 01 09:00:17 crc kubenswrapper[4873]: I1201 09:00:17.525575 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7fc6f67df6-72xpt" podStartSLOduration=2.525550024 podStartE2EDuration="2.525550024s" podCreationTimestamp="2025-12-01 09:00:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:00:17.518887161 +0000 UTC m=+1193.420995700" watchObservedRunningTime="2025-12-01 09:00:17.525550024 +0000 UTC m=+1193.427658563" Dec 01 09:00:17 crc kubenswrapper[4873]: I1201 09:00:17.549917 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.549878659 podStartE2EDuration="3.549878659s" podCreationTimestamp="2025-12-01 09:00:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:00:17.541806462 +0000 UTC m=+1193.443915021" watchObservedRunningTime="2025-12-01 09:00:17.549878659 +0000 UTC m=+1193.451987198" Dec 01 09:00:18 crc kubenswrapper[4873]: I1201 09:00:18.518953 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"15c2d570-e3c5-42e4-b837-b188b167be1f","Type":"ContainerStarted","Data":"095d23cc7ae2b206480a00aec26a4f14779dfbd5a8f3eed34ae5cd04ff093f3e"} Dec 01 09:00:19 crc kubenswrapper[4873]: I1201 09:00:19.077644 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 01 09:00:19 crc kubenswrapper[4873]: I1201 09:00:19.383705 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 01 09:00:19 crc kubenswrapper[4873]: I1201 09:00:19.534823 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"15c2d570-e3c5-42e4-b837-b188b167be1f","Type":"ContainerStarted","Data":"85052f1497d30a5fed0f60503ec9b81af55dd8be20469000257368f913307d7d"} Dec 01 09:00:19 crc kubenswrapper[4873]: I1201 09:00:19.580040 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 09:00:19 crc kubenswrapper[4873]: I1201 09:00:19.792984 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6d97fcdd8f-qmg5n" Dec 01 09:00:19 crc kubenswrapper[4873]: I1201 09:00:19.882746 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b946d459c-csc2t"] Dec 01 09:00:19 crc kubenswrapper[4873]: I1201 09:00:19.883607 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7b946d459c-csc2t" podUID="21098264-17f0-46eb-8531-7df74404e5c7" containerName="dnsmasq-dns" containerID="cri-o://ec850baf9aac29ca65fec4acd31d5cbcdff67cf2b06600c086dd67f67b4840f7" gracePeriod=10 Dec 01 09:00:20 crc kubenswrapper[4873]: I1201 09:00:20.529251 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b946d459c-csc2t" Dec 01 09:00:20 crc kubenswrapper[4873]: I1201 09:00:20.546717 4873 generic.go:334] "Generic (PLEG): container finished" podID="21098264-17f0-46eb-8531-7df74404e5c7" containerID="ec850baf9aac29ca65fec4acd31d5cbcdff67cf2b06600c086dd67f67b4840f7" exitCode=0 Dec 01 09:00:20 crc kubenswrapper[4873]: I1201 09:00:20.547052 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="11731b74-fe99-40c4-9445-f86102242d05" containerName="cinder-scheduler" containerID="cri-o://e77db932d96e0f5c27d59ad43932b7ce4fd4ec9f33e6f1accbc0146e681dcb29" gracePeriod=30 Dec 01 09:00:20 crc kubenswrapper[4873]: I1201 09:00:20.547194 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b946d459c-csc2t" Dec 01 09:00:20 crc kubenswrapper[4873]: I1201 09:00:20.547205 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b946d459c-csc2t" event={"ID":"21098264-17f0-46eb-8531-7df74404e5c7","Type":"ContainerDied","Data":"ec850baf9aac29ca65fec4acd31d5cbcdff67cf2b06600c086dd67f67b4840f7"} Dec 01 09:00:20 crc kubenswrapper[4873]: I1201 09:00:20.547300 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b946d459c-csc2t" event={"ID":"21098264-17f0-46eb-8531-7df74404e5c7","Type":"ContainerDied","Data":"0f59effb39fc254d1509c8e1d2f0dbbda12836bda1f8d041eaf4e461110409f9"} Dec 01 09:00:20 crc kubenswrapper[4873]: I1201 09:00:20.547325 4873 scope.go:117] "RemoveContainer" containerID="ec850baf9aac29ca65fec4acd31d5cbcdff67cf2b06600c086dd67f67b4840f7" Dec 01 09:00:20 crc kubenswrapper[4873]: I1201 09:00:20.547236 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="11731b74-fe99-40c4-9445-f86102242d05" containerName="probe" containerID="cri-o://c4008ccedc96a77868ded5c3f956a834a5990ecf0d660b15d7bb6eb8270f46e7" gracePeriod=30 Dec 01 09:00:20 crc kubenswrapper[4873]: I1201 09:00:20.586379 4873 scope.go:117] "RemoveContainer" containerID="40337c1bdcc463c9337b54bce609af965f1439ed1e6d6a8a3ae9b552e17704d2" Dec 01 09:00:20 crc kubenswrapper[4873]: I1201 09:00:20.629263 4873 scope.go:117] "RemoveContainer" containerID="ec850baf9aac29ca65fec4acd31d5cbcdff67cf2b06600c086dd67f67b4840f7" Dec 01 09:00:20 crc kubenswrapper[4873]: E1201 09:00:20.633153 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec850baf9aac29ca65fec4acd31d5cbcdff67cf2b06600c086dd67f67b4840f7\": container with ID starting with ec850baf9aac29ca65fec4acd31d5cbcdff67cf2b06600c086dd67f67b4840f7 not found: ID does not exist" containerID="ec850baf9aac29ca65fec4acd31d5cbcdff67cf2b06600c086dd67f67b4840f7" Dec 01 09:00:20 crc kubenswrapper[4873]: I1201 09:00:20.633212 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec850baf9aac29ca65fec4acd31d5cbcdff67cf2b06600c086dd67f67b4840f7"} err="failed to get container status \"ec850baf9aac29ca65fec4acd31d5cbcdff67cf2b06600c086dd67f67b4840f7\": rpc error: code = NotFound desc = could not find container \"ec850baf9aac29ca65fec4acd31d5cbcdff67cf2b06600c086dd67f67b4840f7\": container with ID starting with ec850baf9aac29ca65fec4acd31d5cbcdff67cf2b06600c086dd67f67b4840f7 not found: ID does not exist" Dec 01 09:00:20 crc kubenswrapper[4873]: I1201 09:00:20.633255 4873 scope.go:117] "RemoveContainer" containerID="40337c1bdcc463c9337b54bce609af965f1439ed1e6d6a8a3ae9b552e17704d2" Dec 01 09:00:20 crc kubenswrapper[4873]: E1201 09:00:20.637437 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40337c1bdcc463c9337b54bce609af965f1439ed1e6d6a8a3ae9b552e17704d2\": container with ID starting with 40337c1bdcc463c9337b54bce609af965f1439ed1e6d6a8a3ae9b552e17704d2 not found: ID does not exist" containerID="40337c1bdcc463c9337b54bce609af965f1439ed1e6d6a8a3ae9b552e17704d2" Dec 01 09:00:20 crc kubenswrapper[4873]: I1201 09:00:20.637629 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40337c1bdcc463c9337b54bce609af965f1439ed1e6d6a8a3ae9b552e17704d2"} err="failed to get container status \"40337c1bdcc463c9337b54bce609af965f1439ed1e6d6a8a3ae9b552e17704d2\": rpc error: code = NotFound desc = could not find container \"40337c1bdcc463c9337b54bce609af965f1439ed1e6d6a8a3ae9b552e17704d2\": container with ID starting with 40337c1bdcc463c9337b54bce609af965f1439ed1e6d6a8a3ae9b552e17704d2 not found: ID does not exist" Dec 01 09:00:20 crc kubenswrapper[4873]: I1201 09:00:20.710878 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/21098264-17f0-46eb-8531-7df74404e5c7-ovsdbserver-sb\") pod \"21098264-17f0-46eb-8531-7df74404e5c7\" (UID: \"21098264-17f0-46eb-8531-7df74404e5c7\") " Dec 01 09:00:20 crc kubenswrapper[4873]: I1201 09:00:20.711336 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/21098264-17f0-46eb-8531-7df74404e5c7-ovsdbserver-nb\") pod \"21098264-17f0-46eb-8531-7df74404e5c7\" (UID: \"21098264-17f0-46eb-8531-7df74404e5c7\") " Dec 01 09:00:20 crc kubenswrapper[4873]: I1201 09:00:20.711438 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21098264-17f0-46eb-8531-7df74404e5c7-config\") pod \"21098264-17f0-46eb-8531-7df74404e5c7\" (UID: \"21098264-17f0-46eb-8531-7df74404e5c7\") " Dec 01 09:00:20 crc kubenswrapper[4873]: I1201 09:00:20.711661 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/21098264-17f0-46eb-8531-7df74404e5c7-dns-svc\") pod \"21098264-17f0-46eb-8531-7df74404e5c7\" (UID: \"21098264-17f0-46eb-8531-7df74404e5c7\") " Dec 01 09:00:20 crc kubenswrapper[4873]: I1201 09:00:20.711752 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s7wpt\" (UniqueName: \"kubernetes.io/projected/21098264-17f0-46eb-8531-7df74404e5c7-kube-api-access-s7wpt\") pod \"21098264-17f0-46eb-8531-7df74404e5c7\" (UID: \"21098264-17f0-46eb-8531-7df74404e5c7\") " Dec 01 09:00:20 crc kubenswrapper[4873]: I1201 09:00:20.724416 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21098264-17f0-46eb-8531-7df74404e5c7-kube-api-access-s7wpt" (OuterVolumeSpecName: "kube-api-access-s7wpt") pod "21098264-17f0-46eb-8531-7df74404e5c7" (UID: "21098264-17f0-46eb-8531-7df74404e5c7"). InnerVolumeSpecName "kube-api-access-s7wpt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:00:20 crc kubenswrapper[4873]: I1201 09:00:20.799913 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21098264-17f0-46eb-8531-7df74404e5c7-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "21098264-17f0-46eb-8531-7df74404e5c7" (UID: "21098264-17f0-46eb-8531-7df74404e5c7"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:00:20 crc kubenswrapper[4873]: I1201 09:00:20.809304 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21098264-17f0-46eb-8531-7df74404e5c7-config" (OuterVolumeSpecName: "config") pod "21098264-17f0-46eb-8531-7df74404e5c7" (UID: "21098264-17f0-46eb-8531-7df74404e5c7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:00:20 crc kubenswrapper[4873]: I1201 09:00:20.820227 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21098264-17f0-46eb-8531-7df74404e5c7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "21098264-17f0-46eb-8531-7df74404e5c7" (UID: "21098264-17f0-46eb-8531-7df74404e5c7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:00:20 crc kubenswrapper[4873]: I1201 09:00:20.839970 4873 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21098264-17f0-46eb-8531-7df74404e5c7-config\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:20 crc kubenswrapper[4873]: I1201 09:00:20.840035 4873 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/21098264-17f0-46eb-8531-7df74404e5c7-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:20 crc kubenswrapper[4873]: I1201 09:00:20.840074 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s7wpt\" (UniqueName: \"kubernetes.io/projected/21098264-17f0-46eb-8531-7df74404e5c7-kube-api-access-s7wpt\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:20 crc kubenswrapper[4873]: I1201 09:00:20.840096 4873 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/21098264-17f0-46eb-8531-7df74404e5c7-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:20 crc kubenswrapper[4873]: I1201 09:00:20.861321 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21098264-17f0-46eb-8531-7df74404e5c7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "21098264-17f0-46eb-8531-7df74404e5c7" (UID: "21098264-17f0-46eb-8531-7df74404e5c7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:00:20 crc kubenswrapper[4873]: I1201 09:00:20.942852 4873 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/21098264-17f0-46eb-8531-7df74404e5c7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:20 crc kubenswrapper[4873]: I1201 09:00:20.995600 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-56d597fbb8-m5cgw" Dec 01 09:00:21 crc kubenswrapper[4873]: I1201 09:00:21.202831 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b946d459c-csc2t"] Dec 01 09:00:21 crc kubenswrapper[4873]: I1201 09:00:21.221260 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7b946d459c-csc2t"] Dec 01 09:00:21 crc kubenswrapper[4873]: I1201 09:00:21.561886 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"15c2d570-e3c5-42e4-b837-b188b167be1f","Type":"ContainerStarted","Data":"c1c8cc98f705da143058d5c7028ce26dbc4ad601ffb07174032b1e4d2694d7cd"} Dec 01 09:00:21 crc kubenswrapper[4873]: I1201 09:00:21.563804 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 09:00:21 crc kubenswrapper[4873]: I1201 09:00:21.578577 4873 generic.go:334] "Generic (PLEG): container finished" podID="11731b74-fe99-40c4-9445-f86102242d05" containerID="c4008ccedc96a77868ded5c3f956a834a5990ecf0d660b15d7bb6eb8270f46e7" exitCode=0 Dec 01 09:00:21 crc kubenswrapper[4873]: I1201 09:00:21.578646 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"11731b74-fe99-40c4-9445-f86102242d05","Type":"ContainerDied","Data":"c4008ccedc96a77868ded5c3f956a834a5990ecf0d660b15d7bb6eb8270f46e7"} Dec 01 09:00:21 crc kubenswrapper[4873]: I1201 09:00:21.858042 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-64c5c798d6-h87s6" Dec 01 09:00:21 crc kubenswrapper[4873]: I1201 09:00:21.892124 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.28350118 podStartE2EDuration="7.892100461s" podCreationTimestamp="2025-12-01 09:00:14 +0000 UTC" firstStartedPulling="2025-12-01 09:00:15.50907613 +0000 UTC m=+1191.411184669" lastFinishedPulling="2025-12-01 09:00:21.117675411 +0000 UTC m=+1197.019783950" observedRunningTime="2025-12-01 09:00:21.591629922 +0000 UTC m=+1197.493738461" watchObservedRunningTime="2025-12-01 09:00:21.892100461 +0000 UTC m=+1197.794209000" Dec 01 09:00:21 crc kubenswrapper[4873]: I1201 09:00:21.948741 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-64c5c798d6-h87s6" Dec 01 09:00:22 crc kubenswrapper[4873]: I1201 09:00:22.443773 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21098264-17f0-46eb-8531-7df74404e5c7" path="/var/lib/kubelet/pods/21098264-17f0-46eb-8531-7df74404e5c7/volumes" Dec 01 09:00:23 crc kubenswrapper[4873]: I1201 09:00:23.723955 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-f99dddf57-sr27n" Dec 01 09:00:23 crc kubenswrapper[4873]: I1201 09:00:23.797146 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-56d597fbb8-m5cgw"] Dec 01 09:00:23 crc kubenswrapper[4873]: I1201 09:00:23.797392 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-56d597fbb8-m5cgw" podUID="1a53851f-14c9-4577-bb4e-c3a7b6b8c141" containerName="neutron-api" containerID="cri-o://775289bce97005f19b4ef7e8bc7d99f9fad1ebab676f25f563b8efaf92232c61" gracePeriod=30 Dec 01 09:00:23 crc kubenswrapper[4873]: I1201 09:00:23.797779 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-56d597fbb8-m5cgw" podUID="1a53851f-14c9-4577-bb4e-c3a7b6b8c141" containerName="neutron-httpd" containerID="cri-o://86c8b1a2e01ed1ec8a465652759bb8d665cb8b35cb8ff4cd32090ac977656e38" gracePeriod=30 Dec 01 09:00:24 crc kubenswrapper[4873]: I1201 09:00:24.610311 4873 generic.go:334] "Generic (PLEG): container finished" podID="11731b74-fe99-40c4-9445-f86102242d05" containerID="e77db932d96e0f5c27d59ad43932b7ce4fd4ec9f33e6f1accbc0146e681dcb29" exitCode=0 Dec 01 09:00:24 crc kubenswrapper[4873]: I1201 09:00:24.610770 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"11731b74-fe99-40c4-9445-f86102242d05","Type":"ContainerDied","Data":"e77db932d96e0f5c27d59ad43932b7ce4fd4ec9f33e6f1accbc0146e681dcb29"} Dec 01 09:00:24 crc kubenswrapper[4873]: I1201 09:00:24.618436 4873 generic.go:334] "Generic (PLEG): container finished" podID="1a53851f-14c9-4577-bb4e-c3a7b6b8c141" containerID="86c8b1a2e01ed1ec8a465652759bb8d665cb8b35cb8ff4cd32090ac977656e38" exitCode=0 Dec 01 09:00:24 crc kubenswrapper[4873]: I1201 09:00:24.618491 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-56d597fbb8-m5cgw" event={"ID":"1a53851f-14c9-4577-bb4e-c3a7b6b8c141","Type":"ContainerDied","Data":"86c8b1a2e01ed1ec8a465652759bb8d665cb8b35cb8ff4cd32090ac977656e38"} Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.011728 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.074781 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11731b74-fe99-40c4-9445-f86102242d05-combined-ca-bundle\") pod \"11731b74-fe99-40c4-9445-f86102242d05\" (UID: \"11731b74-fe99-40c4-9445-f86102242d05\") " Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.074927 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11731b74-fe99-40c4-9445-f86102242d05-scripts\") pod \"11731b74-fe99-40c4-9445-f86102242d05\" (UID: \"11731b74-fe99-40c4-9445-f86102242d05\") " Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.074970 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11731b74-fe99-40c4-9445-f86102242d05-config-data\") pod \"11731b74-fe99-40c4-9445-f86102242d05\" (UID: \"11731b74-fe99-40c4-9445-f86102242d05\") " Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.075000 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/11731b74-fe99-40c4-9445-f86102242d05-config-data-custom\") pod \"11731b74-fe99-40c4-9445-f86102242d05\" (UID: \"11731b74-fe99-40c4-9445-f86102242d05\") " Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.075044 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/11731b74-fe99-40c4-9445-f86102242d05-etc-machine-id\") pod \"11731b74-fe99-40c4-9445-f86102242d05\" (UID: \"11731b74-fe99-40c4-9445-f86102242d05\") " Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.075081 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n84tf\" (UniqueName: \"kubernetes.io/projected/11731b74-fe99-40c4-9445-f86102242d05-kube-api-access-n84tf\") pod \"11731b74-fe99-40c4-9445-f86102242d05\" (UID: \"11731b74-fe99-40c4-9445-f86102242d05\") " Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.078481 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/11731b74-fe99-40c4-9445-f86102242d05-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "11731b74-fe99-40c4-9445-f86102242d05" (UID: "11731b74-fe99-40c4-9445-f86102242d05"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.084904 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11731b74-fe99-40c4-9445-f86102242d05-kube-api-access-n84tf" (OuterVolumeSpecName: "kube-api-access-n84tf") pod "11731b74-fe99-40c4-9445-f86102242d05" (UID: "11731b74-fe99-40c4-9445-f86102242d05"). InnerVolumeSpecName "kube-api-access-n84tf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.088297 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11731b74-fe99-40c4-9445-f86102242d05-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "11731b74-fe99-40c4-9445-f86102242d05" (UID: "11731b74-fe99-40c4-9445-f86102242d05"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.099490 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11731b74-fe99-40c4-9445-f86102242d05-scripts" (OuterVolumeSpecName: "scripts") pod "11731b74-fe99-40c4-9445-f86102242d05" (UID: "11731b74-fe99-40c4-9445-f86102242d05"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.151157 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11731b74-fe99-40c4-9445-f86102242d05-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "11731b74-fe99-40c4-9445-f86102242d05" (UID: "11731b74-fe99-40c4-9445-f86102242d05"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.176760 4873 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11731b74-fe99-40c4-9445-f86102242d05-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.176826 4873 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/11731b74-fe99-40c4-9445-f86102242d05-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.176836 4873 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/11731b74-fe99-40c4-9445-f86102242d05-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.176844 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n84tf\" (UniqueName: \"kubernetes.io/projected/11731b74-fe99-40c4-9445-f86102242d05-kube-api-access-n84tf\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.176854 4873 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11731b74-fe99-40c4-9445-f86102242d05-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.196255 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11731b74-fe99-40c4-9445-f86102242d05-config-data" (OuterVolumeSpecName: "config-data") pod "11731b74-fe99-40c4-9445-f86102242d05" (UID: "11731b74-fe99-40c4-9445-f86102242d05"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.278639 4873 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11731b74-fe99-40c4-9445-f86102242d05-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.632061 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"11731b74-fe99-40c4-9445-f86102242d05","Type":"ContainerDied","Data":"8358bfa8cb72a469f44ad316593ed5b68a569b1833abfacd2c88aeba8cd0800c"} Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.632163 4873 scope.go:117] "RemoveContainer" containerID="c4008ccedc96a77868ded5c3f956a834a5990ecf0d660b15d7bb6eb8270f46e7" Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.632349 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.669597 4873 scope.go:117] "RemoveContainer" containerID="e77db932d96e0f5c27d59ad43932b7ce4fd4ec9f33e6f1accbc0146e681dcb29" Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.704903 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.717308 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.733094 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 09:00:25 crc kubenswrapper[4873]: E1201 09:00:25.733793 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11731b74-fe99-40c4-9445-f86102242d05" containerName="cinder-scheduler" Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.733819 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="11731b74-fe99-40c4-9445-f86102242d05" containerName="cinder-scheduler" Dec 01 09:00:25 crc kubenswrapper[4873]: E1201 09:00:25.733852 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21098264-17f0-46eb-8531-7df74404e5c7" containerName="dnsmasq-dns" Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.733859 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="21098264-17f0-46eb-8531-7df74404e5c7" containerName="dnsmasq-dns" Dec 01 09:00:25 crc kubenswrapper[4873]: E1201 09:00:25.733869 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11731b74-fe99-40c4-9445-f86102242d05" containerName="probe" Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.733877 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="11731b74-fe99-40c4-9445-f86102242d05" containerName="probe" Dec 01 09:00:25 crc kubenswrapper[4873]: E1201 09:00:25.733899 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21098264-17f0-46eb-8531-7df74404e5c7" containerName="init" Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.733906 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="21098264-17f0-46eb-8531-7df74404e5c7" containerName="init" Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.734184 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="11731b74-fe99-40c4-9445-f86102242d05" containerName="cinder-scheduler" Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.734208 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="21098264-17f0-46eb-8531-7df74404e5c7" containerName="dnsmasq-dns" Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.734230 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="11731b74-fe99-40c4-9445-f86102242d05" containerName="probe" Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.736463 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.740516 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.763610 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.809300 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4c035924-7ef3-4fc1-8652-78317f6b5c70-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"4c035924-7ef3-4fc1-8652-78317f6b5c70\") " pod="openstack/cinder-scheduler-0" Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.809364 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4c035924-7ef3-4fc1-8652-78317f6b5c70-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"4c035924-7ef3-4fc1-8652-78317f6b5c70\") " pod="openstack/cinder-scheduler-0" Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.809401 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c035924-7ef3-4fc1-8652-78317f6b5c70-config-data\") pod \"cinder-scheduler-0\" (UID: \"4c035924-7ef3-4fc1-8652-78317f6b5c70\") " pod="openstack/cinder-scheduler-0" Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.809436 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c035924-7ef3-4fc1-8652-78317f6b5c70-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"4c035924-7ef3-4fc1-8652-78317f6b5c70\") " pod="openstack/cinder-scheduler-0" Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.809469 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cv5rp\" (UniqueName: \"kubernetes.io/projected/4c035924-7ef3-4fc1-8652-78317f6b5c70-kube-api-access-cv5rp\") pod \"cinder-scheduler-0\" (UID: \"4c035924-7ef3-4fc1-8652-78317f6b5c70\") " pod="openstack/cinder-scheduler-0" Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.809494 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c035924-7ef3-4fc1-8652-78317f6b5c70-scripts\") pod \"cinder-scheduler-0\" (UID: \"4c035924-7ef3-4fc1-8652-78317f6b5c70\") " pod="openstack/cinder-scheduler-0" Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.911535 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4c035924-7ef3-4fc1-8652-78317f6b5c70-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"4c035924-7ef3-4fc1-8652-78317f6b5c70\") " pod="openstack/cinder-scheduler-0" Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.912162 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4c035924-7ef3-4fc1-8652-78317f6b5c70-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"4c035924-7ef3-4fc1-8652-78317f6b5c70\") " pod="openstack/cinder-scheduler-0" Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.912286 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4c035924-7ef3-4fc1-8652-78317f6b5c70-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"4c035924-7ef3-4fc1-8652-78317f6b5c70\") " pod="openstack/cinder-scheduler-0" Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.912338 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c035924-7ef3-4fc1-8652-78317f6b5c70-config-data\") pod \"cinder-scheduler-0\" (UID: \"4c035924-7ef3-4fc1-8652-78317f6b5c70\") " pod="openstack/cinder-scheduler-0" Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.912393 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c035924-7ef3-4fc1-8652-78317f6b5c70-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"4c035924-7ef3-4fc1-8652-78317f6b5c70\") " pod="openstack/cinder-scheduler-0" Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.912451 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cv5rp\" (UniqueName: \"kubernetes.io/projected/4c035924-7ef3-4fc1-8652-78317f6b5c70-kube-api-access-cv5rp\") pod \"cinder-scheduler-0\" (UID: \"4c035924-7ef3-4fc1-8652-78317f6b5c70\") " pod="openstack/cinder-scheduler-0" Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.912489 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c035924-7ef3-4fc1-8652-78317f6b5c70-scripts\") pod \"cinder-scheduler-0\" (UID: \"4c035924-7ef3-4fc1-8652-78317f6b5c70\") " pod="openstack/cinder-scheduler-0" Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.923734 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c035924-7ef3-4fc1-8652-78317f6b5c70-scripts\") pod \"cinder-scheduler-0\" (UID: \"4c035924-7ef3-4fc1-8652-78317f6b5c70\") " pod="openstack/cinder-scheduler-0" Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.923763 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c035924-7ef3-4fc1-8652-78317f6b5c70-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"4c035924-7ef3-4fc1-8652-78317f6b5c70\") " pod="openstack/cinder-scheduler-0" Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.924751 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4c035924-7ef3-4fc1-8652-78317f6b5c70-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"4c035924-7ef3-4fc1-8652-78317f6b5c70\") " pod="openstack/cinder-scheduler-0" Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.941246 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c035924-7ef3-4fc1-8652-78317f6b5c70-config-data\") pod \"cinder-scheduler-0\" (UID: \"4c035924-7ef3-4fc1-8652-78317f6b5c70\") " pod="openstack/cinder-scheduler-0" Dec 01 09:00:25 crc kubenswrapper[4873]: I1201 09:00:25.948619 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cv5rp\" (UniqueName: \"kubernetes.io/projected/4c035924-7ef3-4fc1-8652-78317f6b5c70-kube-api-access-cv5rp\") pod \"cinder-scheduler-0\" (UID: \"4c035924-7ef3-4fc1-8652-78317f6b5c70\") " pod="openstack/cinder-scheduler-0" Dec 01 09:00:26 crc kubenswrapper[4873]: I1201 09:00:26.124231 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 09:00:26 crc kubenswrapper[4873]: I1201 09:00:26.457250 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11731b74-fe99-40c4-9445-f86102242d05" path="/var/lib/kubelet/pods/11731b74-fe99-40c4-9445-f86102242d05/volumes" Dec 01 09:00:26 crc kubenswrapper[4873]: I1201 09:00:26.834359 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 09:00:27 crc kubenswrapper[4873]: I1201 09:00:27.671599 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4c035924-7ef3-4fc1-8652-78317f6b5c70","Type":"ContainerStarted","Data":"4226ed088908455a17cb1b1b87aa613402baa035ee9ebda02136112af3992b1a"} Dec 01 09:00:27 crc kubenswrapper[4873]: I1201 09:00:27.792425 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 01 09:00:27 crc kubenswrapper[4873]: I1201 09:00:27.913702 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7d689bd78d-s9rtj" Dec 01 09:00:27 crc kubenswrapper[4873]: I1201 09:00:27.919624 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7fc6f67df6-72xpt" Dec 01 09:00:27 crc kubenswrapper[4873]: I1201 09:00:27.921275 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7d689bd78d-s9rtj" Dec 01 09:00:28 crc kubenswrapper[4873]: I1201 09:00:28.131972 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7fc6f67df6-72xpt" Dec 01 09:00:28 crc kubenswrapper[4873]: I1201 09:00:28.249954 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-64c5c798d6-h87s6"] Dec 01 09:00:28 crc kubenswrapper[4873]: I1201 09:00:28.250289 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-64c5c798d6-h87s6" podUID="c10e1747-9f3e-4fc3-bd55-b83c7a93ea12" containerName="barbican-api-log" containerID="cri-o://9e26d54a7fa344ea38e08b4d1b1bfb07202870fd5bd4d23df69645874876f965" gracePeriod=30 Dec 01 09:00:28 crc kubenswrapper[4873]: I1201 09:00:28.250873 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-64c5c798d6-h87s6" podUID="c10e1747-9f3e-4fc3-bd55-b83c7a93ea12" containerName="barbican-api" containerID="cri-o://6f76fe202e35950a65f0577471da2e00b98b3506205c60cf2b9d9cb456f3eca0" gracePeriod=30 Dec 01 09:00:28 crc kubenswrapper[4873]: I1201 09:00:28.684298 4873 generic.go:334] "Generic (PLEG): container finished" podID="c10e1747-9f3e-4fc3-bd55-b83c7a93ea12" containerID="9e26d54a7fa344ea38e08b4d1b1bfb07202870fd5bd4d23df69645874876f965" exitCode=143 Dec 01 09:00:28 crc kubenswrapper[4873]: I1201 09:00:28.684362 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-64c5c798d6-h87s6" event={"ID":"c10e1747-9f3e-4fc3-bd55-b83c7a93ea12","Type":"ContainerDied","Data":"9e26d54a7fa344ea38e08b4d1b1bfb07202870fd5bd4d23df69645874876f965"} Dec 01 09:00:28 crc kubenswrapper[4873]: I1201 09:00:28.686928 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4c035924-7ef3-4fc1-8652-78317f6b5c70","Type":"ContainerStarted","Data":"13a2a07368306ab12ce9b3b6cbc043d3e11214be6b840d51cdb747f59faa1e7d"} Dec 01 09:00:28 crc kubenswrapper[4873]: I1201 09:00:28.929541 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-58576c9895-sk9tb" Dec 01 09:00:29 crc kubenswrapper[4873]: I1201 09:00:29.698597 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-56d597fbb8-m5cgw" Dec 01 09:00:29 crc kubenswrapper[4873]: I1201 09:00:29.700425 4873 generic.go:334] "Generic (PLEG): container finished" podID="1a53851f-14c9-4577-bb4e-c3a7b6b8c141" containerID="775289bce97005f19b4ef7e8bc7d99f9fad1ebab676f25f563b8efaf92232c61" exitCode=0 Dec 01 09:00:29 crc kubenswrapper[4873]: I1201 09:00:29.700522 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-56d597fbb8-m5cgw" event={"ID":"1a53851f-14c9-4577-bb4e-c3a7b6b8c141","Type":"ContainerDied","Data":"775289bce97005f19b4ef7e8bc7d99f9fad1ebab676f25f563b8efaf92232c61"} Dec 01 09:00:29 crc kubenswrapper[4873]: I1201 09:00:29.700558 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-56d597fbb8-m5cgw" event={"ID":"1a53851f-14c9-4577-bb4e-c3a7b6b8c141","Type":"ContainerDied","Data":"dac62a9725f74be0188bcea687e04bdd8d000362222db48caba4d305358af664"} Dec 01 09:00:29 crc kubenswrapper[4873]: I1201 09:00:29.700581 4873 scope.go:117] "RemoveContainer" containerID="86c8b1a2e01ed1ec8a465652759bb8d665cb8b35cb8ff4cd32090ac977656e38" Dec 01 09:00:29 crc kubenswrapper[4873]: I1201 09:00:29.705815 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4c035924-7ef3-4fc1-8652-78317f6b5c70","Type":"ContainerStarted","Data":"543cb87533846b5b1c886f781c785bb3bb8685173cb1f7dd2840cdd167b31138"} Dec 01 09:00:29 crc kubenswrapper[4873]: I1201 09:00:29.737265 4873 scope.go:117] "RemoveContainer" containerID="775289bce97005f19b4ef7e8bc7d99f9fad1ebab676f25f563b8efaf92232c61" Dec 01 09:00:29 crc kubenswrapper[4873]: I1201 09:00:29.766308 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a53851f-14c9-4577-bb4e-c3a7b6b8c141-combined-ca-bundle\") pod \"1a53851f-14c9-4577-bb4e-c3a7b6b8c141\" (UID: \"1a53851f-14c9-4577-bb4e-c3a7b6b8c141\") " Dec 01 09:00:29 crc kubenswrapper[4873]: I1201 09:00:29.766384 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1a53851f-14c9-4577-bb4e-c3a7b6b8c141-config\") pod \"1a53851f-14c9-4577-bb4e-c3a7b6b8c141\" (UID: \"1a53851f-14c9-4577-bb4e-c3a7b6b8c141\") " Dec 01 09:00:29 crc kubenswrapper[4873]: I1201 09:00:29.766506 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a53851f-14c9-4577-bb4e-c3a7b6b8c141-ovndb-tls-certs\") pod \"1a53851f-14c9-4577-bb4e-c3a7b6b8c141\" (UID: \"1a53851f-14c9-4577-bb4e-c3a7b6b8c141\") " Dec 01 09:00:29 crc kubenswrapper[4873]: I1201 09:00:29.766631 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1a53851f-14c9-4577-bb4e-c3a7b6b8c141-httpd-config\") pod \"1a53851f-14c9-4577-bb4e-c3a7b6b8c141\" (UID: \"1a53851f-14c9-4577-bb4e-c3a7b6b8c141\") " Dec 01 09:00:29 crc kubenswrapper[4873]: I1201 09:00:29.766676 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrrw2\" (UniqueName: \"kubernetes.io/projected/1a53851f-14c9-4577-bb4e-c3a7b6b8c141-kube-api-access-vrrw2\") pod \"1a53851f-14c9-4577-bb4e-c3a7b6b8c141\" (UID: \"1a53851f-14c9-4577-bb4e-c3a7b6b8c141\") " Dec 01 09:00:29 crc kubenswrapper[4873]: I1201 09:00:29.767671 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.767648532 podStartE2EDuration="4.767648532s" podCreationTimestamp="2025-12-01 09:00:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:00:29.753107006 +0000 UTC m=+1205.655215545" watchObservedRunningTime="2025-12-01 09:00:29.767648532 +0000 UTC m=+1205.669757061" Dec 01 09:00:29 crc kubenswrapper[4873]: I1201 09:00:29.778851 4873 scope.go:117] "RemoveContainer" containerID="86c8b1a2e01ed1ec8a465652759bb8d665cb8b35cb8ff4cd32090ac977656e38" Dec 01 09:00:29 crc kubenswrapper[4873]: E1201 09:00:29.779473 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86c8b1a2e01ed1ec8a465652759bb8d665cb8b35cb8ff4cd32090ac977656e38\": container with ID starting with 86c8b1a2e01ed1ec8a465652759bb8d665cb8b35cb8ff4cd32090ac977656e38 not found: ID does not exist" containerID="86c8b1a2e01ed1ec8a465652759bb8d665cb8b35cb8ff4cd32090ac977656e38" Dec 01 09:00:29 crc kubenswrapper[4873]: I1201 09:00:29.779513 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86c8b1a2e01ed1ec8a465652759bb8d665cb8b35cb8ff4cd32090ac977656e38"} err="failed to get container status \"86c8b1a2e01ed1ec8a465652759bb8d665cb8b35cb8ff4cd32090ac977656e38\": rpc error: code = NotFound desc = could not find container \"86c8b1a2e01ed1ec8a465652759bb8d665cb8b35cb8ff4cd32090ac977656e38\": container with ID starting with 86c8b1a2e01ed1ec8a465652759bb8d665cb8b35cb8ff4cd32090ac977656e38 not found: ID does not exist" Dec 01 09:00:29 crc kubenswrapper[4873]: I1201 09:00:29.779550 4873 scope.go:117] "RemoveContainer" containerID="775289bce97005f19b4ef7e8bc7d99f9fad1ebab676f25f563b8efaf92232c61" Dec 01 09:00:29 crc kubenswrapper[4873]: E1201 09:00:29.784670 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"775289bce97005f19b4ef7e8bc7d99f9fad1ebab676f25f563b8efaf92232c61\": container with ID starting with 775289bce97005f19b4ef7e8bc7d99f9fad1ebab676f25f563b8efaf92232c61 not found: ID does not exist" containerID="775289bce97005f19b4ef7e8bc7d99f9fad1ebab676f25f563b8efaf92232c61" Dec 01 09:00:29 crc kubenswrapper[4873]: I1201 09:00:29.784745 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"775289bce97005f19b4ef7e8bc7d99f9fad1ebab676f25f563b8efaf92232c61"} err="failed to get container status \"775289bce97005f19b4ef7e8bc7d99f9fad1ebab676f25f563b8efaf92232c61\": rpc error: code = NotFound desc = could not find container \"775289bce97005f19b4ef7e8bc7d99f9fad1ebab676f25f563b8efaf92232c61\": container with ID starting with 775289bce97005f19b4ef7e8bc7d99f9fad1ebab676f25f563b8efaf92232c61 not found: ID does not exist" Dec 01 09:00:29 crc kubenswrapper[4873]: I1201 09:00:29.788190 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a53851f-14c9-4577-bb4e-c3a7b6b8c141-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "1a53851f-14c9-4577-bb4e-c3a7b6b8c141" (UID: "1a53851f-14c9-4577-bb4e-c3a7b6b8c141"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:29 crc kubenswrapper[4873]: I1201 09:00:29.794101 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a53851f-14c9-4577-bb4e-c3a7b6b8c141-kube-api-access-vrrw2" (OuterVolumeSpecName: "kube-api-access-vrrw2") pod "1a53851f-14c9-4577-bb4e-c3a7b6b8c141" (UID: "1a53851f-14c9-4577-bb4e-c3a7b6b8c141"). InnerVolumeSpecName "kube-api-access-vrrw2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:00:29 crc kubenswrapper[4873]: I1201 09:00:29.850988 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a53851f-14c9-4577-bb4e-c3a7b6b8c141-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1a53851f-14c9-4577-bb4e-c3a7b6b8c141" (UID: "1a53851f-14c9-4577-bb4e-c3a7b6b8c141"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:29 crc kubenswrapper[4873]: I1201 09:00:29.868300 4873 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1a53851f-14c9-4577-bb4e-c3a7b6b8c141-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:29 crc kubenswrapper[4873]: I1201 09:00:29.868342 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrrw2\" (UniqueName: \"kubernetes.io/projected/1a53851f-14c9-4577-bb4e-c3a7b6b8c141-kube-api-access-vrrw2\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:29 crc kubenswrapper[4873]: I1201 09:00:29.868353 4873 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a53851f-14c9-4577-bb4e-c3a7b6b8c141-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:29 crc kubenswrapper[4873]: I1201 09:00:29.896078 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a53851f-14c9-4577-bb4e-c3a7b6b8c141-config" (OuterVolumeSpecName: "config") pod "1a53851f-14c9-4577-bb4e-c3a7b6b8c141" (UID: "1a53851f-14c9-4577-bb4e-c3a7b6b8c141"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:29 crc kubenswrapper[4873]: I1201 09:00:29.911419 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a53851f-14c9-4577-bb4e-c3a7b6b8c141-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "1a53851f-14c9-4577-bb4e-c3a7b6b8c141" (UID: "1a53851f-14c9-4577-bb4e-c3a7b6b8c141"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:29 crc kubenswrapper[4873]: I1201 09:00:29.969704 4873 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/1a53851f-14c9-4577-bb4e-c3a7b6b8c141-config\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:29 crc kubenswrapper[4873]: I1201 09:00:29.970275 4873 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a53851f-14c9-4577-bb4e-c3a7b6b8c141-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:30 crc kubenswrapper[4873]: I1201 09:00:30.719842 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-56d597fbb8-m5cgw" Dec 01 09:00:30 crc kubenswrapper[4873]: I1201 09:00:30.748380 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-56d597fbb8-m5cgw"] Dec 01 09:00:30 crc kubenswrapper[4873]: I1201 09:00:30.755186 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-56d597fbb8-m5cgw"] Dec 01 09:00:31 crc kubenswrapper[4873]: I1201 09:00:31.125340 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 01 09:00:31 crc kubenswrapper[4873]: I1201 09:00:31.737048 4873 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-64c5c798d6-h87s6" podUID="c10e1747-9f3e-4fc3-bd55-b83c7a93ea12" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.149:9311/healthcheck\": read tcp 10.217.0.2:54026->10.217.0.149:9311: read: connection reset by peer" Dec 01 09:00:31 crc kubenswrapper[4873]: I1201 09:00:31.737867 4873 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-64c5c798d6-h87s6" podUID="c10e1747-9f3e-4fc3-bd55-b83c7a93ea12" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.149:9311/healthcheck\": read tcp 10.217.0.2:54032->10.217.0.149:9311: read: connection reset by peer" Dec 01 09:00:32 crc kubenswrapper[4873]: E1201 09:00:32.108640 4873 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc10e1747_9f3e_4fc3_bd55_b83c7a93ea12.slice/crio-conmon-6f76fe202e35950a65f0577471da2e00b98b3506205c60cf2b9d9cb456f3eca0.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a53851f_14c9_4577_bb4e_c3a7b6b8c141.slice/crio-conmon-775289bce97005f19b4ef7e8bc7d99f9fad1ebab676f25f563b8efaf92232c61.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc10e1747_9f3e_4fc3_bd55_b83c7a93ea12.slice/crio-6f76fe202e35950a65f0577471da2e00b98b3506205c60cf2b9d9cb456f3eca0.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a53851f_14c9_4577_bb4e_c3a7b6b8c141.slice/crio-775289bce97005f19b4ef7e8bc7d99f9fad1ebab676f25f563b8efaf92232c61.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a53851f_14c9_4577_bb4e_c3a7b6b8c141.slice/crio-dac62a9725f74be0188bcea687e04bdd8d000362222db48caba4d305358af664\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a53851f_14c9_4577_bb4e_c3a7b6b8c141.slice\": RecentStats: unable to find data in memory cache]" Dec 01 09:00:32 crc kubenswrapper[4873]: I1201 09:00:32.355064 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-64c5c798d6-h87s6" Dec 01 09:00:32 crc kubenswrapper[4873]: I1201 09:00:32.441826 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a53851f-14c9-4577-bb4e-c3a7b6b8c141" path="/var/lib/kubelet/pods/1a53851f-14c9-4577-bb4e-c3a7b6b8c141/volumes" Dec 01 09:00:32 crc kubenswrapper[4873]: I1201 09:00:32.452725 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c10e1747-9f3e-4fc3-bd55-b83c7a93ea12-logs\") pod \"c10e1747-9f3e-4fc3-bd55-b83c7a93ea12\" (UID: \"c10e1747-9f3e-4fc3-bd55-b83c7a93ea12\") " Dec 01 09:00:32 crc kubenswrapper[4873]: I1201 09:00:32.452914 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c10e1747-9f3e-4fc3-bd55-b83c7a93ea12-config-data\") pod \"c10e1747-9f3e-4fc3-bd55-b83c7a93ea12\" (UID: \"c10e1747-9f3e-4fc3-bd55-b83c7a93ea12\") " Dec 01 09:00:32 crc kubenswrapper[4873]: I1201 09:00:32.452980 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mj4f8\" (UniqueName: \"kubernetes.io/projected/c10e1747-9f3e-4fc3-bd55-b83c7a93ea12-kube-api-access-mj4f8\") pod \"c10e1747-9f3e-4fc3-bd55-b83c7a93ea12\" (UID: \"c10e1747-9f3e-4fc3-bd55-b83c7a93ea12\") " Dec 01 09:00:32 crc kubenswrapper[4873]: I1201 09:00:32.453056 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c10e1747-9f3e-4fc3-bd55-b83c7a93ea12-config-data-custom\") pod \"c10e1747-9f3e-4fc3-bd55-b83c7a93ea12\" (UID: \"c10e1747-9f3e-4fc3-bd55-b83c7a93ea12\") " Dec 01 09:00:32 crc kubenswrapper[4873]: I1201 09:00:32.453090 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c10e1747-9f3e-4fc3-bd55-b83c7a93ea12-combined-ca-bundle\") pod \"c10e1747-9f3e-4fc3-bd55-b83c7a93ea12\" (UID: \"c10e1747-9f3e-4fc3-bd55-b83c7a93ea12\") " Dec 01 09:00:32 crc kubenswrapper[4873]: I1201 09:00:32.453643 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c10e1747-9f3e-4fc3-bd55-b83c7a93ea12-logs" (OuterVolumeSpecName: "logs") pod "c10e1747-9f3e-4fc3-bd55-b83c7a93ea12" (UID: "c10e1747-9f3e-4fc3-bd55-b83c7a93ea12"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:00:32 crc kubenswrapper[4873]: I1201 09:00:32.469347 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c10e1747-9f3e-4fc3-bd55-b83c7a93ea12-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "c10e1747-9f3e-4fc3-bd55-b83c7a93ea12" (UID: "c10e1747-9f3e-4fc3-bd55-b83c7a93ea12"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:32 crc kubenswrapper[4873]: I1201 09:00:32.481373 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c10e1747-9f3e-4fc3-bd55-b83c7a93ea12-kube-api-access-mj4f8" (OuterVolumeSpecName: "kube-api-access-mj4f8") pod "c10e1747-9f3e-4fc3-bd55-b83c7a93ea12" (UID: "c10e1747-9f3e-4fc3-bd55-b83c7a93ea12"). InnerVolumeSpecName "kube-api-access-mj4f8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:00:32 crc kubenswrapper[4873]: I1201 09:00:32.529202 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c10e1747-9f3e-4fc3-bd55-b83c7a93ea12-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c10e1747-9f3e-4fc3-bd55-b83c7a93ea12" (UID: "c10e1747-9f3e-4fc3-bd55-b83c7a93ea12"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:32 crc kubenswrapper[4873]: I1201 09:00:32.543459 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c10e1747-9f3e-4fc3-bd55-b83c7a93ea12-config-data" (OuterVolumeSpecName: "config-data") pod "c10e1747-9f3e-4fc3-bd55-b83c7a93ea12" (UID: "c10e1747-9f3e-4fc3-bd55-b83c7a93ea12"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:32 crc kubenswrapper[4873]: I1201 09:00:32.556219 4873 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c10e1747-9f3e-4fc3-bd55-b83c7a93ea12-logs\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:32 crc kubenswrapper[4873]: I1201 09:00:32.556272 4873 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c10e1747-9f3e-4fc3-bd55-b83c7a93ea12-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:32 crc kubenswrapper[4873]: I1201 09:00:32.556288 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mj4f8\" (UniqueName: \"kubernetes.io/projected/c10e1747-9f3e-4fc3-bd55-b83c7a93ea12-kube-api-access-mj4f8\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:32 crc kubenswrapper[4873]: I1201 09:00:32.556305 4873 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c10e1747-9f3e-4fc3-bd55-b83c7a93ea12-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:32 crc kubenswrapper[4873]: I1201 09:00:32.556316 4873 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c10e1747-9f3e-4fc3-bd55-b83c7a93ea12-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:32 crc kubenswrapper[4873]: I1201 09:00:32.741066 4873 generic.go:334] "Generic (PLEG): container finished" podID="c10e1747-9f3e-4fc3-bd55-b83c7a93ea12" containerID="6f76fe202e35950a65f0577471da2e00b98b3506205c60cf2b9d9cb456f3eca0" exitCode=0 Dec 01 09:00:32 crc kubenswrapper[4873]: I1201 09:00:32.741119 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-64c5c798d6-h87s6" event={"ID":"c10e1747-9f3e-4fc3-bd55-b83c7a93ea12","Type":"ContainerDied","Data":"6f76fe202e35950a65f0577471da2e00b98b3506205c60cf2b9d9cb456f3eca0"} Dec 01 09:00:32 crc kubenswrapper[4873]: I1201 09:00:32.741152 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-64c5c798d6-h87s6" event={"ID":"c10e1747-9f3e-4fc3-bd55-b83c7a93ea12","Type":"ContainerDied","Data":"721d8e182e7f528d1281b938b4c11f54863a9778f23f7c3b9f7de84e7e36cb59"} Dec 01 09:00:32 crc kubenswrapper[4873]: I1201 09:00:32.741174 4873 scope.go:117] "RemoveContainer" containerID="6f76fe202e35950a65f0577471da2e00b98b3506205c60cf2b9d9cb456f3eca0" Dec 01 09:00:32 crc kubenswrapper[4873]: I1201 09:00:32.741303 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-64c5c798d6-h87s6" Dec 01 09:00:32 crc kubenswrapper[4873]: I1201 09:00:32.785440 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-64c5c798d6-h87s6"] Dec 01 09:00:32 crc kubenswrapper[4873]: I1201 09:00:32.790200 4873 scope.go:117] "RemoveContainer" containerID="9e26d54a7fa344ea38e08b4d1b1bfb07202870fd5bd4d23df69645874876f965" Dec 01 09:00:32 crc kubenswrapper[4873]: I1201 09:00:32.798855 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-64c5c798d6-h87s6"] Dec 01 09:00:32 crc kubenswrapper[4873]: I1201 09:00:32.820898 4873 scope.go:117] "RemoveContainer" containerID="6f76fe202e35950a65f0577471da2e00b98b3506205c60cf2b9d9cb456f3eca0" Dec 01 09:00:32 crc kubenswrapper[4873]: E1201 09:00:32.821804 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f76fe202e35950a65f0577471da2e00b98b3506205c60cf2b9d9cb456f3eca0\": container with ID starting with 6f76fe202e35950a65f0577471da2e00b98b3506205c60cf2b9d9cb456f3eca0 not found: ID does not exist" containerID="6f76fe202e35950a65f0577471da2e00b98b3506205c60cf2b9d9cb456f3eca0" Dec 01 09:00:32 crc kubenswrapper[4873]: I1201 09:00:32.821890 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f76fe202e35950a65f0577471da2e00b98b3506205c60cf2b9d9cb456f3eca0"} err="failed to get container status \"6f76fe202e35950a65f0577471da2e00b98b3506205c60cf2b9d9cb456f3eca0\": rpc error: code = NotFound desc = could not find container \"6f76fe202e35950a65f0577471da2e00b98b3506205c60cf2b9d9cb456f3eca0\": container with ID starting with 6f76fe202e35950a65f0577471da2e00b98b3506205c60cf2b9d9cb456f3eca0 not found: ID does not exist" Dec 01 09:00:32 crc kubenswrapper[4873]: I1201 09:00:32.821953 4873 scope.go:117] "RemoveContainer" containerID="9e26d54a7fa344ea38e08b4d1b1bfb07202870fd5bd4d23df69645874876f965" Dec 01 09:00:32 crc kubenswrapper[4873]: E1201 09:00:32.822505 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e26d54a7fa344ea38e08b4d1b1bfb07202870fd5bd4d23df69645874876f965\": container with ID starting with 9e26d54a7fa344ea38e08b4d1b1bfb07202870fd5bd4d23df69645874876f965 not found: ID does not exist" containerID="9e26d54a7fa344ea38e08b4d1b1bfb07202870fd5bd4d23df69645874876f965" Dec 01 09:00:32 crc kubenswrapper[4873]: I1201 09:00:32.822538 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e26d54a7fa344ea38e08b4d1b1bfb07202870fd5bd4d23df69645874876f965"} err="failed to get container status \"9e26d54a7fa344ea38e08b4d1b1bfb07202870fd5bd4d23df69645874876f965\": rpc error: code = NotFound desc = could not find container \"9e26d54a7fa344ea38e08b4d1b1bfb07202870fd5bd4d23df69645874876f965\": container with ID starting with 9e26d54a7fa344ea38e08b4d1b1bfb07202870fd5bd4d23df69645874876f965 not found: ID does not exist" Dec 01 09:00:33 crc kubenswrapper[4873]: I1201 09:00:33.635375 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 01 09:00:33 crc kubenswrapper[4873]: E1201 09:00:33.641623 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c10e1747-9f3e-4fc3-bd55-b83c7a93ea12" containerName="barbican-api-log" Dec 01 09:00:33 crc kubenswrapper[4873]: I1201 09:00:33.641666 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="c10e1747-9f3e-4fc3-bd55-b83c7a93ea12" containerName="barbican-api-log" Dec 01 09:00:33 crc kubenswrapper[4873]: E1201 09:00:33.641682 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a53851f-14c9-4577-bb4e-c3a7b6b8c141" containerName="neutron-httpd" Dec 01 09:00:33 crc kubenswrapper[4873]: I1201 09:00:33.641690 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a53851f-14c9-4577-bb4e-c3a7b6b8c141" containerName="neutron-httpd" Dec 01 09:00:33 crc kubenswrapper[4873]: E1201 09:00:33.641704 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c10e1747-9f3e-4fc3-bd55-b83c7a93ea12" containerName="barbican-api" Dec 01 09:00:33 crc kubenswrapper[4873]: I1201 09:00:33.641711 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="c10e1747-9f3e-4fc3-bd55-b83c7a93ea12" containerName="barbican-api" Dec 01 09:00:33 crc kubenswrapper[4873]: E1201 09:00:33.641725 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a53851f-14c9-4577-bb4e-c3a7b6b8c141" containerName="neutron-api" Dec 01 09:00:33 crc kubenswrapper[4873]: I1201 09:00:33.641732 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a53851f-14c9-4577-bb4e-c3a7b6b8c141" containerName="neutron-api" Dec 01 09:00:33 crc kubenswrapper[4873]: I1201 09:00:33.641977 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a53851f-14c9-4577-bb4e-c3a7b6b8c141" containerName="neutron-api" Dec 01 09:00:33 crc kubenswrapper[4873]: I1201 09:00:33.641997 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="c10e1747-9f3e-4fc3-bd55-b83c7a93ea12" containerName="barbican-api-log" Dec 01 09:00:33 crc kubenswrapper[4873]: I1201 09:00:33.642034 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a53851f-14c9-4577-bb4e-c3a7b6b8c141" containerName="neutron-httpd" Dec 01 09:00:33 crc kubenswrapper[4873]: I1201 09:00:33.642049 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="c10e1747-9f3e-4fc3-bd55-b83c7a93ea12" containerName="barbican-api" Dec 01 09:00:33 crc kubenswrapper[4873]: I1201 09:00:33.643106 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 09:00:33 crc kubenswrapper[4873]: I1201 09:00:33.644286 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 01 09:00:33 crc kubenswrapper[4873]: I1201 09:00:33.645845 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-6wblt" Dec 01 09:00:33 crc kubenswrapper[4873]: I1201 09:00:33.646711 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 01 09:00:33 crc kubenswrapper[4873]: I1201 09:00:33.652280 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 01 09:00:33 crc kubenswrapper[4873]: I1201 09:00:33.687490 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhn67\" (UniqueName: \"kubernetes.io/projected/43a97c60-b43e-4896-9592-d41863b1c203-kube-api-access-dhn67\") pod \"openstackclient\" (UID: \"43a97c60-b43e-4896-9592-d41863b1c203\") " pod="openstack/openstackclient" Dec 01 09:00:33 crc kubenswrapper[4873]: I1201 09:00:33.687527 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/43a97c60-b43e-4896-9592-d41863b1c203-openstack-config-secret\") pod \"openstackclient\" (UID: \"43a97c60-b43e-4896-9592-d41863b1c203\") " pod="openstack/openstackclient" Dec 01 09:00:33 crc kubenswrapper[4873]: I1201 09:00:33.687688 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/43a97c60-b43e-4896-9592-d41863b1c203-openstack-config\") pod \"openstackclient\" (UID: \"43a97c60-b43e-4896-9592-d41863b1c203\") " pod="openstack/openstackclient" Dec 01 09:00:33 crc kubenswrapper[4873]: I1201 09:00:33.687745 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43a97c60-b43e-4896-9592-d41863b1c203-combined-ca-bundle\") pod \"openstackclient\" (UID: \"43a97c60-b43e-4896-9592-d41863b1c203\") " pod="openstack/openstackclient" Dec 01 09:00:33 crc kubenswrapper[4873]: I1201 09:00:33.789041 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhn67\" (UniqueName: \"kubernetes.io/projected/43a97c60-b43e-4896-9592-d41863b1c203-kube-api-access-dhn67\") pod \"openstackclient\" (UID: \"43a97c60-b43e-4896-9592-d41863b1c203\") " pod="openstack/openstackclient" Dec 01 09:00:33 crc kubenswrapper[4873]: I1201 09:00:33.789088 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/43a97c60-b43e-4896-9592-d41863b1c203-openstack-config-secret\") pod \"openstackclient\" (UID: \"43a97c60-b43e-4896-9592-d41863b1c203\") " pod="openstack/openstackclient" Dec 01 09:00:33 crc kubenswrapper[4873]: I1201 09:00:33.789191 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/43a97c60-b43e-4896-9592-d41863b1c203-openstack-config\") pod \"openstackclient\" (UID: \"43a97c60-b43e-4896-9592-d41863b1c203\") " pod="openstack/openstackclient" Dec 01 09:00:33 crc kubenswrapper[4873]: I1201 09:00:33.789251 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43a97c60-b43e-4896-9592-d41863b1c203-combined-ca-bundle\") pod \"openstackclient\" (UID: \"43a97c60-b43e-4896-9592-d41863b1c203\") " pod="openstack/openstackclient" Dec 01 09:00:33 crc kubenswrapper[4873]: I1201 09:00:33.790437 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/43a97c60-b43e-4896-9592-d41863b1c203-openstack-config\") pod \"openstackclient\" (UID: \"43a97c60-b43e-4896-9592-d41863b1c203\") " pod="openstack/openstackclient" Dec 01 09:00:33 crc kubenswrapper[4873]: I1201 09:00:33.795571 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43a97c60-b43e-4896-9592-d41863b1c203-combined-ca-bundle\") pod \"openstackclient\" (UID: \"43a97c60-b43e-4896-9592-d41863b1c203\") " pod="openstack/openstackclient" Dec 01 09:00:33 crc kubenswrapper[4873]: I1201 09:00:33.795793 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/43a97c60-b43e-4896-9592-d41863b1c203-openstack-config-secret\") pod \"openstackclient\" (UID: \"43a97c60-b43e-4896-9592-d41863b1c203\") " pod="openstack/openstackclient" Dec 01 09:00:33 crc kubenswrapper[4873]: I1201 09:00:33.809820 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhn67\" (UniqueName: \"kubernetes.io/projected/43a97c60-b43e-4896-9592-d41863b1c203-kube-api-access-dhn67\") pod \"openstackclient\" (UID: \"43a97c60-b43e-4896-9592-d41863b1c203\") " pod="openstack/openstackclient" Dec 01 09:00:33 crc kubenswrapper[4873]: I1201 09:00:33.993138 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 09:00:34 crc kubenswrapper[4873]: I1201 09:00:34.442443 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c10e1747-9f3e-4fc3-bd55-b83c7a93ea12" path="/var/lib/kubelet/pods/c10e1747-9f3e-4fc3-bd55-b83c7a93ea12/volumes" Dec 01 09:00:34 crc kubenswrapper[4873]: I1201 09:00:34.491154 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 01 09:00:34 crc kubenswrapper[4873]: I1201 09:00:34.766154 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"43a97c60-b43e-4896-9592-d41863b1c203","Type":"ContainerStarted","Data":"ea7dffb458201193157090684685a6bdd2e40f5824eb44933625aeab70a6e4a5"} Dec 01 09:00:36 crc kubenswrapper[4873]: I1201 09:00:36.408601 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 01 09:00:43 crc kubenswrapper[4873]: I1201 09:00:43.734298 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 09:00:43 crc kubenswrapper[4873]: I1201 09:00:43.736327 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="15c2d570-e3c5-42e4-b837-b188b167be1f" containerName="ceilometer-central-agent" containerID="cri-o://f21f06a3b000822f1f73e94cdc69477bc1636249d7d3ec510cf2b5a38321bc7b" gracePeriod=30 Dec 01 09:00:43 crc kubenswrapper[4873]: I1201 09:00:43.736480 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="15c2d570-e3c5-42e4-b837-b188b167be1f" containerName="ceilometer-notification-agent" containerID="cri-o://095d23cc7ae2b206480a00aec26a4f14779dfbd5a8f3eed34ae5cd04ff093f3e" gracePeriod=30 Dec 01 09:00:43 crc kubenswrapper[4873]: I1201 09:00:43.736445 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="15c2d570-e3c5-42e4-b837-b188b167be1f" containerName="sg-core" containerID="cri-o://85052f1497d30a5fed0f60503ec9b81af55dd8be20469000257368f913307d7d" gracePeriod=30 Dec 01 09:00:43 crc kubenswrapper[4873]: I1201 09:00:43.736488 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="15c2d570-e3c5-42e4-b837-b188b167be1f" containerName="proxy-httpd" containerID="cri-o://c1c8cc98f705da143058d5c7028ce26dbc4ad601ffb07174032b1e4d2694d7cd" gracePeriod=30 Dec 01 09:00:43 crc kubenswrapper[4873]: I1201 09:00:43.749577 4873 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="15c2d570-e3c5-42e4-b837-b188b167be1f" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Dec 01 09:00:43 crc kubenswrapper[4873]: I1201 09:00:43.875417 4873 generic.go:334] "Generic (PLEG): container finished" podID="15c2d570-e3c5-42e4-b837-b188b167be1f" containerID="85052f1497d30a5fed0f60503ec9b81af55dd8be20469000257368f913307d7d" exitCode=2 Dec 01 09:00:43 crc kubenswrapper[4873]: I1201 09:00:43.875848 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"15c2d570-e3c5-42e4-b837-b188b167be1f","Type":"ContainerDied","Data":"85052f1497d30a5fed0f60503ec9b81af55dd8be20469000257368f913307d7d"} Dec 01 09:00:44 crc kubenswrapper[4873]: I1201 09:00:44.890892 4873 generic.go:334] "Generic (PLEG): container finished" podID="15c2d570-e3c5-42e4-b837-b188b167be1f" containerID="c1c8cc98f705da143058d5c7028ce26dbc4ad601ffb07174032b1e4d2694d7cd" exitCode=0 Dec 01 09:00:44 crc kubenswrapper[4873]: I1201 09:00:44.891574 4873 generic.go:334] "Generic (PLEG): container finished" podID="15c2d570-e3c5-42e4-b837-b188b167be1f" containerID="f21f06a3b000822f1f73e94cdc69477bc1636249d7d3ec510cf2b5a38321bc7b" exitCode=0 Dec 01 09:00:44 crc kubenswrapper[4873]: I1201 09:00:44.891637 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"15c2d570-e3c5-42e4-b837-b188b167be1f","Type":"ContainerDied","Data":"c1c8cc98f705da143058d5c7028ce26dbc4ad601ffb07174032b1e4d2694d7cd"} Dec 01 09:00:44 crc kubenswrapper[4873]: I1201 09:00:44.891671 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"15c2d570-e3c5-42e4-b837-b188b167be1f","Type":"ContainerDied","Data":"f21f06a3b000822f1f73e94cdc69477bc1636249d7d3ec510cf2b5a38321bc7b"} Dec 01 09:00:44 crc kubenswrapper[4873]: I1201 09:00:44.895122 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"43a97c60-b43e-4896-9592-d41863b1c203","Type":"ContainerStarted","Data":"c1d808d26e3d6644cfdbb3fe680c42ae23dc5cdc56a53287474ee5535f2ede34"} Dec 01 09:00:44 crc kubenswrapper[4873]: I1201 09:00:44.913816 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.659943891 podStartE2EDuration="11.913773678s" podCreationTimestamp="2025-12-01 09:00:33 +0000 UTC" firstStartedPulling="2025-12-01 09:00:34.498276202 +0000 UTC m=+1210.400384741" lastFinishedPulling="2025-12-01 09:00:43.752105979 +0000 UTC m=+1219.654214528" observedRunningTime="2025-12-01 09:00:44.911693117 +0000 UTC m=+1220.813801656" watchObservedRunningTime="2025-12-01 09:00:44.913773678 +0000 UTC m=+1220.815882217" Dec 01 09:00:44 crc kubenswrapper[4873]: I1201 09:00:44.915608 4873 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="15c2d570-e3c5-42e4-b837-b188b167be1f" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.150:3000/\": dial tcp 10.217.0.150:3000: connect: connection refused" Dec 01 09:00:46 crc kubenswrapper[4873]: I1201 09:00:46.651097 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 09:00:46 crc kubenswrapper[4873]: I1201 09:00:46.739004 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15c2d570-e3c5-42e4-b837-b188b167be1f-combined-ca-bundle\") pod \"15c2d570-e3c5-42e4-b837-b188b167be1f\" (UID: \"15c2d570-e3c5-42e4-b837-b188b167be1f\") " Dec 01 09:00:46 crc kubenswrapper[4873]: I1201 09:00:46.739092 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/15c2d570-e3c5-42e4-b837-b188b167be1f-log-httpd\") pod \"15c2d570-e3c5-42e4-b837-b188b167be1f\" (UID: \"15c2d570-e3c5-42e4-b837-b188b167be1f\") " Dec 01 09:00:46 crc kubenswrapper[4873]: I1201 09:00:46.739123 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15c2d570-e3c5-42e4-b837-b188b167be1f-scripts\") pod \"15c2d570-e3c5-42e4-b837-b188b167be1f\" (UID: \"15c2d570-e3c5-42e4-b837-b188b167be1f\") " Dec 01 09:00:46 crc kubenswrapper[4873]: I1201 09:00:46.739199 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/15c2d570-e3c5-42e4-b837-b188b167be1f-sg-core-conf-yaml\") pod \"15c2d570-e3c5-42e4-b837-b188b167be1f\" (UID: \"15c2d570-e3c5-42e4-b837-b188b167be1f\") " Dec 01 09:00:46 crc kubenswrapper[4873]: I1201 09:00:46.739234 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15c2d570-e3c5-42e4-b837-b188b167be1f-config-data\") pod \"15c2d570-e3c5-42e4-b837-b188b167be1f\" (UID: \"15c2d570-e3c5-42e4-b837-b188b167be1f\") " Dec 01 09:00:46 crc kubenswrapper[4873]: I1201 09:00:46.739323 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p8ct2\" (UniqueName: \"kubernetes.io/projected/15c2d570-e3c5-42e4-b837-b188b167be1f-kube-api-access-p8ct2\") pod \"15c2d570-e3c5-42e4-b837-b188b167be1f\" (UID: \"15c2d570-e3c5-42e4-b837-b188b167be1f\") " Dec 01 09:00:46 crc kubenswrapper[4873]: I1201 09:00:46.739376 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/15c2d570-e3c5-42e4-b837-b188b167be1f-run-httpd\") pod \"15c2d570-e3c5-42e4-b837-b188b167be1f\" (UID: \"15c2d570-e3c5-42e4-b837-b188b167be1f\") " Dec 01 09:00:46 crc kubenswrapper[4873]: I1201 09:00:46.740378 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15c2d570-e3c5-42e4-b837-b188b167be1f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "15c2d570-e3c5-42e4-b837-b188b167be1f" (UID: "15c2d570-e3c5-42e4-b837-b188b167be1f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:00:46 crc kubenswrapper[4873]: I1201 09:00:46.741399 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15c2d570-e3c5-42e4-b837-b188b167be1f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "15c2d570-e3c5-42e4-b837-b188b167be1f" (UID: "15c2d570-e3c5-42e4-b837-b188b167be1f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:00:46 crc kubenswrapper[4873]: I1201 09:00:46.759690 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15c2d570-e3c5-42e4-b837-b188b167be1f-kube-api-access-p8ct2" (OuterVolumeSpecName: "kube-api-access-p8ct2") pod "15c2d570-e3c5-42e4-b837-b188b167be1f" (UID: "15c2d570-e3c5-42e4-b837-b188b167be1f"). InnerVolumeSpecName "kube-api-access-p8ct2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:00:46 crc kubenswrapper[4873]: I1201 09:00:46.759806 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15c2d570-e3c5-42e4-b837-b188b167be1f-scripts" (OuterVolumeSpecName: "scripts") pod "15c2d570-e3c5-42e4-b837-b188b167be1f" (UID: "15c2d570-e3c5-42e4-b837-b188b167be1f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:46 crc kubenswrapper[4873]: I1201 09:00:46.781309 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15c2d570-e3c5-42e4-b837-b188b167be1f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "15c2d570-e3c5-42e4-b837-b188b167be1f" (UID: "15c2d570-e3c5-42e4-b837-b188b167be1f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:46 crc kubenswrapper[4873]: I1201 09:00:46.829381 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15c2d570-e3c5-42e4-b837-b188b167be1f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "15c2d570-e3c5-42e4-b837-b188b167be1f" (UID: "15c2d570-e3c5-42e4-b837-b188b167be1f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:46 crc kubenswrapper[4873]: I1201 09:00:46.841584 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p8ct2\" (UniqueName: \"kubernetes.io/projected/15c2d570-e3c5-42e4-b837-b188b167be1f-kube-api-access-p8ct2\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:46 crc kubenswrapper[4873]: I1201 09:00:46.841621 4873 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/15c2d570-e3c5-42e4-b837-b188b167be1f-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:46 crc kubenswrapper[4873]: I1201 09:00:46.841631 4873 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15c2d570-e3c5-42e4-b837-b188b167be1f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:46 crc kubenswrapper[4873]: I1201 09:00:46.841641 4873 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/15c2d570-e3c5-42e4-b837-b188b167be1f-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:46 crc kubenswrapper[4873]: I1201 09:00:46.841649 4873 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15c2d570-e3c5-42e4-b837-b188b167be1f-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:46 crc kubenswrapper[4873]: I1201 09:00:46.841659 4873 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/15c2d570-e3c5-42e4-b837-b188b167be1f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:46 crc kubenswrapper[4873]: I1201 09:00:46.856488 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15c2d570-e3c5-42e4-b837-b188b167be1f-config-data" (OuterVolumeSpecName: "config-data") pod "15c2d570-e3c5-42e4-b837-b188b167be1f" (UID: "15c2d570-e3c5-42e4-b837-b188b167be1f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:46 crc kubenswrapper[4873]: I1201 09:00:46.919471 4873 generic.go:334] "Generic (PLEG): container finished" podID="15c2d570-e3c5-42e4-b837-b188b167be1f" containerID="095d23cc7ae2b206480a00aec26a4f14779dfbd5a8f3eed34ae5cd04ff093f3e" exitCode=0 Dec 01 09:00:46 crc kubenswrapper[4873]: I1201 09:00:46.919537 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"15c2d570-e3c5-42e4-b837-b188b167be1f","Type":"ContainerDied","Data":"095d23cc7ae2b206480a00aec26a4f14779dfbd5a8f3eed34ae5cd04ff093f3e"} Dec 01 09:00:46 crc kubenswrapper[4873]: I1201 09:00:46.919599 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"15c2d570-e3c5-42e4-b837-b188b167be1f","Type":"ContainerDied","Data":"f2209bbb291eace67e6c2a39ebb72b7ba687e697d507435485f29499349fecb0"} Dec 01 09:00:46 crc kubenswrapper[4873]: I1201 09:00:46.919631 4873 scope.go:117] "RemoveContainer" containerID="c1c8cc98f705da143058d5c7028ce26dbc4ad601ffb07174032b1e4d2694d7cd" Dec 01 09:00:46 crc kubenswrapper[4873]: I1201 09:00:46.919764 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 09:00:46 crc kubenswrapper[4873]: I1201 09:00:46.943589 4873 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15c2d570-e3c5-42e4-b837-b188b167be1f-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:46 crc kubenswrapper[4873]: I1201 09:00:46.947915 4873 scope.go:117] "RemoveContainer" containerID="85052f1497d30a5fed0f60503ec9b81af55dd8be20469000257368f913307d7d" Dec 01 09:00:46 crc kubenswrapper[4873]: I1201 09:00:46.977548 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 09:00:46 crc kubenswrapper[4873]: I1201 09:00:46.977766 4873 scope.go:117] "RemoveContainer" containerID="095d23cc7ae2b206480a00aec26a4f14779dfbd5a8f3eed34ae5cd04ff093f3e" Dec 01 09:00:46 crc kubenswrapper[4873]: I1201 09:00:46.998483 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 09:00:47 crc kubenswrapper[4873]: I1201 09:00:47.007717 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 09:00:47 crc kubenswrapper[4873]: E1201 09:00:47.008151 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15c2d570-e3c5-42e4-b837-b188b167be1f" containerName="sg-core" Dec 01 09:00:47 crc kubenswrapper[4873]: I1201 09:00:47.008169 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="15c2d570-e3c5-42e4-b837-b188b167be1f" containerName="sg-core" Dec 01 09:00:47 crc kubenswrapper[4873]: E1201 09:00:47.008184 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15c2d570-e3c5-42e4-b837-b188b167be1f" containerName="proxy-httpd" Dec 01 09:00:47 crc kubenswrapper[4873]: I1201 09:00:47.008191 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="15c2d570-e3c5-42e4-b837-b188b167be1f" containerName="proxy-httpd" Dec 01 09:00:47 crc kubenswrapper[4873]: E1201 09:00:47.008207 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15c2d570-e3c5-42e4-b837-b188b167be1f" containerName="ceilometer-central-agent" Dec 01 09:00:47 crc kubenswrapper[4873]: I1201 09:00:47.008214 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="15c2d570-e3c5-42e4-b837-b188b167be1f" containerName="ceilometer-central-agent" Dec 01 09:00:47 crc kubenswrapper[4873]: E1201 09:00:47.008240 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15c2d570-e3c5-42e4-b837-b188b167be1f" containerName="ceilometer-notification-agent" Dec 01 09:00:47 crc kubenswrapper[4873]: I1201 09:00:47.008247 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="15c2d570-e3c5-42e4-b837-b188b167be1f" containerName="ceilometer-notification-agent" Dec 01 09:00:47 crc kubenswrapper[4873]: I1201 09:00:47.008451 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="15c2d570-e3c5-42e4-b837-b188b167be1f" containerName="ceilometer-central-agent" Dec 01 09:00:47 crc kubenswrapper[4873]: I1201 09:00:47.008468 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="15c2d570-e3c5-42e4-b837-b188b167be1f" containerName="ceilometer-notification-agent" Dec 01 09:00:47 crc kubenswrapper[4873]: I1201 09:00:47.008479 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="15c2d570-e3c5-42e4-b837-b188b167be1f" containerName="proxy-httpd" Dec 01 09:00:47 crc kubenswrapper[4873]: I1201 09:00:47.008490 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="15c2d570-e3c5-42e4-b837-b188b167be1f" containerName="sg-core" Dec 01 09:00:47 crc kubenswrapper[4873]: I1201 09:00:47.010193 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 09:00:47 crc kubenswrapper[4873]: I1201 09:00:47.015514 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 09:00:47 crc kubenswrapper[4873]: I1201 09:00:47.015885 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 09:00:47 crc kubenswrapper[4873]: I1201 09:00:47.030294 4873 scope.go:117] "RemoveContainer" containerID="f21f06a3b000822f1f73e94cdc69477bc1636249d7d3ec510cf2b5a38321bc7b" Dec 01 09:00:47 crc kubenswrapper[4873]: I1201 09:00:47.055595 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 09:00:47 crc kubenswrapper[4873]: I1201 09:00:47.077767 4873 scope.go:117] "RemoveContainer" containerID="c1c8cc98f705da143058d5c7028ce26dbc4ad601ffb07174032b1e4d2694d7cd" Dec 01 09:00:47 crc kubenswrapper[4873]: E1201 09:00:47.078421 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1c8cc98f705da143058d5c7028ce26dbc4ad601ffb07174032b1e4d2694d7cd\": container with ID starting with c1c8cc98f705da143058d5c7028ce26dbc4ad601ffb07174032b1e4d2694d7cd not found: ID does not exist" containerID="c1c8cc98f705da143058d5c7028ce26dbc4ad601ffb07174032b1e4d2694d7cd" Dec 01 09:00:47 crc kubenswrapper[4873]: I1201 09:00:47.078464 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1c8cc98f705da143058d5c7028ce26dbc4ad601ffb07174032b1e4d2694d7cd"} err="failed to get container status \"c1c8cc98f705da143058d5c7028ce26dbc4ad601ffb07174032b1e4d2694d7cd\": rpc error: code = NotFound desc = could not find container \"c1c8cc98f705da143058d5c7028ce26dbc4ad601ffb07174032b1e4d2694d7cd\": container with ID starting with c1c8cc98f705da143058d5c7028ce26dbc4ad601ffb07174032b1e4d2694d7cd not found: ID does not exist" Dec 01 09:00:47 crc kubenswrapper[4873]: I1201 09:00:47.078500 4873 scope.go:117] "RemoveContainer" containerID="85052f1497d30a5fed0f60503ec9b81af55dd8be20469000257368f913307d7d" Dec 01 09:00:47 crc kubenswrapper[4873]: E1201 09:00:47.078987 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85052f1497d30a5fed0f60503ec9b81af55dd8be20469000257368f913307d7d\": container with ID starting with 85052f1497d30a5fed0f60503ec9b81af55dd8be20469000257368f913307d7d not found: ID does not exist" containerID="85052f1497d30a5fed0f60503ec9b81af55dd8be20469000257368f913307d7d" Dec 01 09:00:47 crc kubenswrapper[4873]: I1201 09:00:47.079030 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85052f1497d30a5fed0f60503ec9b81af55dd8be20469000257368f913307d7d"} err="failed to get container status \"85052f1497d30a5fed0f60503ec9b81af55dd8be20469000257368f913307d7d\": rpc error: code = NotFound desc = could not find container \"85052f1497d30a5fed0f60503ec9b81af55dd8be20469000257368f913307d7d\": container with ID starting with 85052f1497d30a5fed0f60503ec9b81af55dd8be20469000257368f913307d7d not found: ID does not exist" Dec 01 09:00:47 crc kubenswrapper[4873]: I1201 09:00:47.079052 4873 scope.go:117] "RemoveContainer" containerID="095d23cc7ae2b206480a00aec26a4f14779dfbd5a8f3eed34ae5cd04ff093f3e" Dec 01 09:00:47 crc kubenswrapper[4873]: E1201 09:00:47.079326 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"095d23cc7ae2b206480a00aec26a4f14779dfbd5a8f3eed34ae5cd04ff093f3e\": container with ID starting with 095d23cc7ae2b206480a00aec26a4f14779dfbd5a8f3eed34ae5cd04ff093f3e not found: ID does not exist" containerID="095d23cc7ae2b206480a00aec26a4f14779dfbd5a8f3eed34ae5cd04ff093f3e" Dec 01 09:00:47 crc kubenswrapper[4873]: I1201 09:00:47.079358 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"095d23cc7ae2b206480a00aec26a4f14779dfbd5a8f3eed34ae5cd04ff093f3e"} err="failed to get container status \"095d23cc7ae2b206480a00aec26a4f14779dfbd5a8f3eed34ae5cd04ff093f3e\": rpc error: code = NotFound desc = could not find container \"095d23cc7ae2b206480a00aec26a4f14779dfbd5a8f3eed34ae5cd04ff093f3e\": container with ID starting with 095d23cc7ae2b206480a00aec26a4f14779dfbd5a8f3eed34ae5cd04ff093f3e not found: ID does not exist" Dec 01 09:00:47 crc kubenswrapper[4873]: I1201 09:00:47.079378 4873 scope.go:117] "RemoveContainer" containerID="f21f06a3b000822f1f73e94cdc69477bc1636249d7d3ec510cf2b5a38321bc7b" Dec 01 09:00:47 crc kubenswrapper[4873]: E1201 09:00:47.079644 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f21f06a3b000822f1f73e94cdc69477bc1636249d7d3ec510cf2b5a38321bc7b\": container with ID starting with f21f06a3b000822f1f73e94cdc69477bc1636249d7d3ec510cf2b5a38321bc7b not found: ID does not exist" containerID="f21f06a3b000822f1f73e94cdc69477bc1636249d7d3ec510cf2b5a38321bc7b" Dec 01 09:00:47 crc kubenswrapper[4873]: I1201 09:00:47.079673 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f21f06a3b000822f1f73e94cdc69477bc1636249d7d3ec510cf2b5a38321bc7b"} err="failed to get container status \"f21f06a3b000822f1f73e94cdc69477bc1636249d7d3ec510cf2b5a38321bc7b\": rpc error: code = NotFound desc = could not find container \"f21f06a3b000822f1f73e94cdc69477bc1636249d7d3ec510cf2b5a38321bc7b\": container with ID starting with f21f06a3b000822f1f73e94cdc69477bc1636249d7d3ec510cf2b5a38321bc7b not found: ID does not exist" Dec 01 09:00:47 crc kubenswrapper[4873]: I1201 09:00:47.149402 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b918b2a5-a989-44af-918c-850b1da98aa6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b918b2a5-a989-44af-918c-850b1da98aa6\") " pod="openstack/ceilometer-0" Dec 01 09:00:47 crc kubenswrapper[4873]: I1201 09:00:47.149474 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b918b2a5-a989-44af-918c-850b1da98aa6-scripts\") pod \"ceilometer-0\" (UID: \"b918b2a5-a989-44af-918c-850b1da98aa6\") " pod="openstack/ceilometer-0" Dec 01 09:00:47 crc kubenswrapper[4873]: I1201 09:00:47.149576 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b918b2a5-a989-44af-918c-850b1da98aa6-run-httpd\") pod \"ceilometer-0\" (UID: \"b918b2a5-a989-44af-918c-850b1da98aa6\") " pod="openstack/ceilometer-0" Dec 01 09:00:47 crc kubenswrapper[4873]: I1201 09:00:47.149684 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b918b2a5-a989-44af-918c-850b1da98aa6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b918b2a5-a989-44af-918c-850b1da98aa6\") " pod="openstack/ceilometer-0" Dec 01 09:00:47 crc kubenswrapper[4873]: I1201 09:00:47.149748 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b918b2a5-a989-44af-918c-850b1da98aa6-config-data\") pod \"ceilometer-0\" (UID: \"b918b2a5-a989-44af-918c-850b1da98aa6\") " pod="openstack/ceilometer-0" Dec 01 09:00:47 crc kubenswrapper[4873]: I1201 09:00:47.149999 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksjl5\" (UniqueName: \"kubernetes.io/projected/b918b2a5-a989-44af-918c-850b1da98aa6-kube-api-access-ksjl5\") pod \"ceilometer-0\" (UID: \"b918b2a5-a989-44af-918c-850b1da98aa6\") " pod="openstack/ceilometer-0" Dec 01 09:00:47 crc kubenswrapper[4873]: I1201 09:00:47.150048 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b918b2a5-a989-44af-918c-850b1da98aa6-log-httpd\") pod \"ceilometer-0\" (UID: \"b918b2a5-a989-44af-918c-850b1da98aa6\") " pod="openstack/ceilometer-0" Dec 01 09:00:47 crc kubenswrapper[4873]: I1201 09:00:47.251918 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b918b2a5-a989-44af-918c-850b1da98aa6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b918b2a5-a989-44af-918c-850b1da98aa6\") " pod="openstack/ceilometer-0" Dec 01 09:00:47 crc kubenswrapper[4873]: I1201 09:00:47.251976 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b918b2a5-a989-44af-918c-850b1da98aa6-scripts\") pod \"ceilometer-0\" (UID: \"b918b2a5-a989-44af-918c-850b1da98aa6\") " pod="openstack/ceilometer-0" Dec 01 09:00:47 crc kubenswrapper[4873]: I1201 09:00:47.252007 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b918b2a5-a989-44af-918c-850b1da98aa6-run-httpd\") pod \"ceilometer-0\" (UID: \"b918b2a5-a989-44af-918c-850b1da98aa6\") " pod="openstack/ceilometer-0" Dec 01 09:00:47 crc kubenswrapper[4873]: I1201 09:00:47.252057 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b918b2a5-a989-44af-918c-850b1da98aa6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b918b2a5-a989-44af-918c-850b1da98aa6\") " pod="openstack/ceilometer-0" Dec 01 09:00:47 crc kubenswrapper[4873]: I1201 09:00:47.252097 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b918b2a5-a989-44af-918c-850b1da98aa6-config-data\") pod \"ceilometer-0\" (UID: \"b918b2a5-a989-44af-918c-850b1da98aa6\") " pod="openstack/ceilometer-0" Dec 01 09:00:47 crc kubenswrapper[4873]: I1201 09:00:47.252169 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksjl5\" (UniqueName: \"kubernetes.io/projected/b918b2a5-a989-44af-918c-850b1da98aa6-kube-api-access-ksjl5\") pod \"ceilometer-0\" (UID: \"b918b2a5-a989-44af-918c-850b1da98aa6\") " pod="openstack/ceilometer-0" Dec 01 09:00:47 crc kubenswrapper[4873]: I1201 09:00:47.252193 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b918b2a5-a989-44af-918c-850b1da98aa6-log-httpd\") pod \"ceilometer-0\" (UID: \"b918b2a5-a989-44af-918c-850b1da98aa6\") " pod="openstack/ceilometer-0" Dec 01 09:00:47 crc kubenswrapper[4873]: I1201 09:00:47.252730 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b918b2a5-a989-44af-918c-850b1da98aa6-run-httpd\") pod \"ceilometer-0\" (UID: \"b918b2a5-a989-44af-918c-850b1da98aa6\") " pod="openstack/ceilometer-0" Dec 01 09:00:47 crc kubenswrapper[4873]: I1201 09:00:47.252885 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b918b2a5-a989-44af-918c-850b1da98aa6-log-httpd\") pod \"ceilometer-0\" (UID: \"b918b2a5-a989-44af-918c-850b1da98aa6\") " pod="openstack/ceilometer-0" Dec 01 09:00:47 crc kubenswrapper[4873]: I1201 09:00:47.256120 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b918b2a5-a989-44af-918c-850b1da98aa6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b918b2a5-a989-44af-918c-850b1da98aa6\") " pod="openstack/ceilometer-0" Dec 01 09:00:47 crc kubenswrapper[4873]: I1201 09:00:47.257025 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b918b2a5-a989-44af-918c-850b1da98aa6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b918b2a5-a989-44af-918c-850b1da98aa6\") " pod="openstack/ceilometer-0" Dec 01 09:00:47 crc kubenswrapper[4873]: I1201 09:00:47.257687 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b918b2a5-a989-44af-918c-850b1da98aa6-scripts\") pod \"ceilometer-0\" (UID: \"b918b2a5-a989-44af-918c-850b1da98aa6\") " pod="openstack/ceilometer-0" Dec 01 09:00:47 crc kubenswrapper[4873]: I1201 09:00:47.259094 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b918b2a5-a989-44af-918c-850b1da98aa6-config-data\") pod \"ceilometer-0\" (UID: \"b918b2a5-a989-44af-918c-850b1da98aa6\") " pod="openstack/ceilometer-0" Dec 01 09:00:47 crc kubenswrapper[4873]: I1201 09:00:47.275977 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksjl5\" (UniqueName: \"kubernetes.io/projected/b918b2a5-a989-44af-918c-850b1da98aa6-kube-api-access-ksjl5\") pod \"ceilometer-0\" (UID: \"b918b2a5-a989-44af-918c-850b1da98aa6\") " pod="openstack/ceilometer-0" Dec 01 09:00:47 crc kubenswrapper[4873]: I1201 09:00:47.336465 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 09:00:47 crc kubenswrapper[4873]: I1201 09:00:47.831868 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 09:00:47 crc kubenswrapper[4873]: I1201 09:00:47.930679 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b918b2a5-a989-44af-918c-850b1da98aa6","Type":"ContainerStarted","Data":"6217417204f3c1ead004fb62de272090b3d61efe6754a38a917f86b53737383c"} Dec 01 09:00:48 crc kubenswrapper[4873]: I1201 09:00:48.025303 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 09:00:48 crc kubenswrapper[4873]: I1201 09:00:48.448569 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15c2d570-e3c5-42e4-b837-b188b167be1f" path="/var/lib/kubelet/pods/15c2d570-e3c5-42e4-b837-b188b167be1f/volumes" Dec 01 09:00:48 crc kubenswrapper[4873]: I1201 09:00:48.943995 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b918b2a5-a989-44af-918c-850b1da98aa6","Type":"ContainerStarted","Data":"c8399d5ddeaad30fb97eb909507984286dce7ccfac74ca7f7ef408cfa836976a"} Dec 01 09:00:49 crc kubenswrapper[4873]: I1201 09:00:49.955752 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b918b2a5-a989-44af-918c-850b1da98aa6","Type":"ContainerStarted","Data":"5320235a23db62440b92a97af8ee2daef5081e6943bb7cb26244b3824d6d215a"} Dec 01 09:00:50 crc kubenswrapper[4873]: I1201 09:00:50.968830 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b918b2a5-a989-44af-918c-850b1da98aa6","Type":"ContainerStarted","Data":"cbcee76f688c973d2de72244c4863f1145ed52772ea13fd3025d5fd2bdd8b557"} Dec 01 09:00:52 crc kubenswrapper[4873]: I1201 09:00:52.994620 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b918b2a5-a989-44af-918c-850b1da98aa6","Type":"ContainerStarted","Data":"f1c8856b9b92a20b6d58ccb9e8add6d1e249ba9f5b36cef739db859b609430ed"} Dec 01 09:00:52 crc kubenswrapper[4873]: I1201 09:00:52.995699 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 09:00:52 crc kubenswrapper[4873]: I1201 09:00:52.994950 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b918b2a5-a989-44af-918c-850b1da98aa6" containerName="ceilometer-notification-agent" containerID="cri-o://5320235a23db62440b92a97af8ee2daef5081e6943bb7cb26244b3824d6d215a" gracePeriod=30 Dec 01 09:00:52 crc kubenswrapper[4873]: I1201 09:00:52.994877 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b918b2a5-a989-44af-918c-850b1da98aa6" containerName="ceilometer-central-agent" containerID="cri-o://c8399d5ddeaad30fb97eb909507984286dce7ccfac74ca7f7ef408cfa836976a" gracePeriod=30 Dec 01 09:00:52 crc kubenswrapper[4873]: I1201 09:00:52.995070 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b918b2a5-a989-44af-918c-850b1da98aa6" containerName="proxy-httpd" containerID="cri-o://f1c8856b9b92a20b6d58ccb9e8add6d1e249ba9f5b36cef739db859b609430ed" gracePeriod=30 Dec 01 09:00:52 crc kubenswrapper[4873]: I1201 09:00:52.994967 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b918b2a5-a989-44af-918c-850b1da98aa6" containerName="sg-core" containerID="cri-o://cbcee76f688c973d2de72244c4863f1145ed52772ea13fd3025d5fd2bdd8b557" gracePeriod=30 Dec 01 09:00:53 crc kubenswrapper[4873]: I1201 09:00:53.030512 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.072585234 podStartE2EDuration="7.030485947s" podCreationTimestamp="2025-12-01 09:00:46 +0000 UTC" firstStartedPulling="2025-12-01 09:00:47.843129607 +0000 UTC m=+1223.745238146" lastFinishedPulling="2025-12-01 09:00:51.80103031 +0000 UTC m=+1227.703138859" observedRunningTime="2025-12-01 09:00:53.028924959 +0000 UTC m=+1228.931033508" watchObservedRunningTime="2025-12-01 09:00:53.030485947 +0000 UTC m=+1228.932594486" Dec 01 09:00:53 crc kubenswrapper[4873]: I1201 09:00:53.991071 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.013588 4873 generic.go:334] "Generic (PLEG): container finished" podID="b918b2a5-a989-44af-918c-850b1da98aa6" containerID="f1c8856b9b92a20b6d58ccb9e8add6d1e249ba9f5b36cef739db859b609430ed" exitCode=0 Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.013636 4873 generic.go:334] "Generic (PLEG): container finished" podID="b918b2a5-a989-44af-918c-850b1da98aa6" containerID="cbcee76f688c973d2de72244c4863f1145ed52772ea13fd3025d5fd2bdd8b557" exitCode=2 Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.013647 4873 generic.go:334] "Generic (PLEG): container finished" podID="b918b2a5-a989-44af-918c-850b1da98aa6" containerID="5320235a23db62440b92a97af8ee2daef5081e6943bb7cb26244b3824d6d215a" exitCode=0 Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.013655 4873 generic.go:334] "Generic (PLEG): container finished" podID="b918b2a5-a989-44af-918c-850b1da98aa6" containerID="c8399d5ddeaad30fb97eb909507984286dce7ccfac74ca7f7ef408cfa836976a" exitCode=0 Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.013670 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.013732 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b918b2a5-a989-44af-918c-850b1da98aa6","Type":"ContainerDied","Data":"f1c8856b9b92a20b6d58ccb9e8add6d1e249ba9f5b36cef739db859b609430ed"} Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.013837 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b918b2a5-a989-44af-918c-850b1da98aa6","Type":"ContainerDied","Data":"cbcee76f688c973d2de72244c4863f1145ed52772ea13fd3025d5fd2bdd8b557"} Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.013853 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b918b2a5-a989-44af-918c-850b1da98aa6","Type":"ContainerDied","Data":"5320235a23db62440b92a97af8ee2daef5081e6943bb7cb26244b3824d6d215a"} Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.013866 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b918b2a5-a989-44af-918c-850b1da98aa6","Type":"ContainerDied","Data":"c8399d5ddeaad30fb97eb909507984286dce7ccfac74ca7f7ef408cfa836976a"} Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.013879 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b918b2a5-a989-44af-918c-850b1da98aa6","Type":"ContainerDied","Data":"6217417204f3c1ead004fb62de272090b3d61efe6754a38a917f86b53737383c"} Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.013945 4873 scope.go:117] "RemoveContainer" containerID="f1c8856b9b92a20b6d58ccb9e8add6d1e249ba9f5b36cef739db859b609430ed" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.033652 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b918b2a5-a989-44af-918c-850b1da98aa6-scripts\") pod \"b918b2a5-a989-44af-918c-850b1da98aa6\" (UID: \"b918b2a5-a989-44af-918c-850b1da98aa6\") " Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.033700 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b918b2a5-a989-44af-918c-850b1da98aa6-combined-ca-bundle\") pod \"b918b2a5-a989-44af-918c-850b1da98aa6\" (UID: \"b918b2a5-a989-44af-918c-850b1da98aa6\") " Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.033810 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b918b2a5-a989-44af-918c-850b1da98aa6-log-httpd\") pod \"b918b2a5-a989-44af-918c-850b1da98aa6\" (UID: \"b918b2a5-a989-44af-918c-850b1da98aa6\") " Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.033907 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b918b2a5-a989-44af-918c-850b1da98aa6-run-httpd\") pod \"b918b2a5-a989-44af-918c-850b1da98aa6\" (UID: \"b918b2a5-a989-44af-918c-850b1da98aa6\") " Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.033946 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b918b2a5-a989-44af-918c-850b1da98aa6-config-data\") pod \"b918b2a5-a989-44af-918c-850b1da98aa6\" (UID: \"b918b2a5-a989-44af-918c-850b1da98aa6\") " Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.034002 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ksjl5\" (UniqueName: \"kubernetes.io/projected/b918b2a5-a989-44af-918c-850b1da98aa6-kube-api-access-ksjl5\") pod \"b918b2a5-a989-44af-918c-850b1da98aa6\" (UID: \"b918b2a5-a989-44af-918c-850b1da98aa6\") " Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.034049 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b918b2a5-a989-44af-918c-850b1da98aa6-sg-core-conf-yaml\") pod \"b918b2a5-a989-44af-918c-850b1da98aa6\" (UID: \"b918b2a5-a989-44af-918c-850b1da98aa6\") " Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.034813 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b918b2a5-a989-44af-918c-850b1da98aa6-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b918b2a5-a989-44af-918c-850b1da98aa6" (UID: "b918b2a5-a989-44af-918c-850b1da98aa6"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.034808 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b918b2a5-a989-44af-918c-850b1da98aa6-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b918b2a5-a989-44af-918c-850b1da98aa6" (UID: "b918b2a5-a989-44af-918c-850b1da98aa6"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.075815 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b918b2a5-a989-44af-918c-850b1da98aa6-scripts" (OuterVolumeSpecName: "scripts") pod "b918b2a5-a989-44af-918c-850b1da98aa6" (UID: "b918b2a5-a989-44af-918c-850b1da98aa6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.079264 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b918b2a5-a989-44af-918c-850b1da98aa6-kube-api-access-ksjl5" (OuterVolumeSpecName: "kube-api-access-ksjl5") pod "b918b2a5-a989-44af-918c-850b1da98aa6" (UID: "b918b2a5-a989-44af-918c-850b1da98aa6"). InnerVolumeSpecName "kube-api-access-ksjl5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.085858 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b918b2a5-a989-44af-918c-850b1da98aa6-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b918b2a5-a989-44af-918c-850b1da98aa6" (UID: "b918b2a5-a989-44af-918c-850b1da98aa6"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.087447 4873 scope.go:117] "RemoveContainer" containerID="cbcee76f688c973d2de72244c4863f1145ed52772ea13fd3025d5fd2bdd8b557" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.113516 4873 scope.go:117] "RemoveContainer" containerID="5320235a23db62440b92a97af8ee2daef5081e6943bb7cb26244b3824d6d215a" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.137474 4873 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b918b2a5-a989-44af-918c-850b1da98aa6-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.137510 4873 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b918b2a5-a989-44af-918c-850b1da98aa6-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.137526 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ksjl5\" (UniqueName: \"kubernetes.io/projected/b918b2a5-a989-44af-918c-850b1da98aa6-kube-api-access-ksjl5\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.137541 4873 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b918b2a5-a989-44af-918c-850b1da98aa6-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.137553 4873 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b918b2a5-a989-44af-918c-850b1da98aa6-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.148984 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b918b2a5-a989-44af-918c-850b1da98aa6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b918b2a5-a989-44af-918c-850b1da98aa6" (UID: "b918b2a5-a989-44af-918c-850b1da98aa6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.169091 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b918b2a5-a989-44af-918c-850b1da98aa6-config-data" (OuterVolumeSpecName: "config-data") pod "b918b2a5-a989-44af-918c-850b1da98aa6" (UID: "b918b2a5-a989-44af-918c-850b1da98aa6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.229994 4873 scope.go:117] "RemoveContainer" containerID="c8399d5ddeaad30fb97eb909507984286dce7ccfac74ca7f7ef408cfa836976a" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.238407 4873 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b918b2a5-a989-44af-918c-850b1da98aa6-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.238449 4873 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b918b2a5-a989-44af-918c-850b1da98aa6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.249631 4873 scope.go:117] "RemoveContainer" containerID="f1c8856b9b92a20b6d58ccb9e8add6d1e249ba9f5b36cef739db859b609430ed" Dec 01 09:00:54 crc kubenswrapper[4873]: E1201 09:00:54.250175 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1c8856b9b92a20b6d58ccb9e8add6d1e249ba9f5b36cef739db859b609430ed\": container with ID starting with f1c8856b9b92a20b6d58ccb9e8add6d1e249ba9f5b36cef739db859b609430ed not found: ID does not exist" containerID="f1c8856b9b92a20b6d58ccb9e8add6d1e249ba9f5b36cef739db859b609430ed" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.250237 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1c8856b9b92a20b6d58ccb9e8add6d1e249ba9f5b36cef739db859b609430ed"} err="failed to get container status \"f1c8856b9b92a20b6d58ccb9e8add6d1e249ba9f5b36cef739db859b609430ed\": rpc error: code = NotFound desc = could not find container \"f1c8856b9b92a20b6d58ccb9e8add6d1e249ba9f5b36cef739db859b609430ed\": container with ID starting with f1c8856b9b92a20b6d58ccb9e8add6d1e249ba9f5b36cef739db859b609430ed not found: ID does not exist" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.250281 4873 scope.go:117] "RemoveContainer" containerID="cbcee76f688c973d2de72244c4863f1145ed52772ea13fd3025d5fd2bdd8b557" Dec 01 09:00:54 crc kubenswrapper[4873]: E1201 09:00:54.250997 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cbcee76f688c973d2de72244c4863f1145ed52772ea13fd3025d5fd2bdd8b557\": container with ID starting with cbcee76f688c973d2de72244c4863f1145ed52772ea13fd3025d5fd2bdd8b557 not found: ID does not exist" containerID="cbcee76f688c973d2de72244c4863f1145ed52772ea13fd3025d5fd2bdd8b557" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.251038 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cbcee76f688c973d2de72244c4863f1145ed52772ea13fd3025d5fd2bdd8b557"} err="failed to get container status \"cbcee76f688c973d2de72244c4863f1145ed52772ea13fd3025d5fd2bdd8b557\": rpc error: code = NotFound desc = could not find container \"cbcee76f688c973d2de72244c4863f1145ed52772ea13fd3025d5fd2bdd8b557\": container with ID starting with cbcee76f688c973d2de72244c4863f1145ed52772ea13fd3025d5fd2bdd8b557 not found: ID does not exist" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.251057 4873 scope.go:117] "RemoveContainer" containerID="5320235a23db62440b92a97af8ee2daef5081e6943bb7cb26244b3824d6d215a" Dec 01 09:00:54 crc kubenswrapper[4873]: E1201 09:00:54.251305 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5320235a23db62440b92a97af8ee2daef5081e6943bb7cb26244b3824d6d215a\": container with ID starting with 5320235a23db62440b92a97af8ee2daef5081e6943bb7cb26244b3824d6d215a not found: ID does not exist" containerID="5320235a23db62440b92a97af8ee2daef5081e6943bb7cb26244b3824d6d215a" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.251332 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5320235a23db62440b92a97af8ee2daef5081e6943bb7cb26244b3824d6d215a"} err="failed to get container status \"5320235a23db62440b92a97af8ee2daef5081e6943bb7cb26244b3824d6d215a\": rpc error: code = NotFound desc = could not find container \"5320235a23db62440b92a97af8ee2daef5081e6943bb7cb26244b3824d6d215a\": container with ID starting with 5320235a23db62440b92a97af8ee2daef5081e6943bb7cb26244b3824d6d215a not found: ID does not exist" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.251347 4873 scope.go:117] "RemoveContainer" containerID="c8399d5ddeaad30fb97eb909507984286dce7ccfac74ca7f7ef408cfa836976a" Dec 01 09:00:54 crc kubenswrapper[4873]: E1201 09:00:54.251928 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8399d5ddeaad30fb97eb909507984286dce7ccfac74ca7f7ef408cfa836976a\": container with ID starting with c8399d5ddeaad30fb97eb909507984286dce7ccfac74ca7f7ef408cfa836976a not found: ID does not exist" containerID="c8399d5ddeaad30fb97eb909507984286dce7ccfac74ca7f7ef408cfa836976a" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.251955 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8399d5ddeaad30fb97eb909507984286dce7ccfac74ca7f7ef408cfa836976a"} err="failed to get container status \"c8399d5ddeaad30fb97eb909507984286dce7ccfac74ca7f7ef408cfa836976a\": rpc error: code = NotFound desc = could not find container \"c8399d5ddeaad30fb97eb909507984286dce7ccfac74ca7f7ef408cfa836976a\": container with ID starting with c8399d5ddeaad30fb97eb909507984286dce7ccfac74ca7f7ef408cfa836976a not found: ID does not exist" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.251970 4873 scope.go:117] "RemoveContainer" containerID="f1c8856b9b92a20b6d58ccb9e8add6d1e249ba9f5b36cef739db859b609430ed" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.252431 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1c8856b9b92a20b6d58ccb9e8add6d1e249ba9f5b36cef739db859b609430ed"} err="failed to get container status \"f1c8856b9b92a20b6d58ccb9e8add6d1e249ba9f5b36cef739db859b609430ed\": rpc error: code = NotFound desc = could not find container \"f1c8856b9b92a20b6d58ccb9e8add6d1e249ba9f5b36cef739db859b609430ed\": container with ID starting with f1c8856b9b92a20b6d58ccb9e8add6d1e249ba9f5b36cef739db859b609430ed not found: ID does not exist" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.252451 4873 scope.go:117] "RemoveContainer" containerID="cbcee76f688c973d2de72244c4863f1145ed52772ea13fd3025d5fd2bdd8b557" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.252792 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cbcee76f688c973d2de72244c4863f1145ed52772ea13fd3025d5fd2bdd8b557"} err="failed to get container status \"cbcee76f688c973d2de72244c4863f1145ed52772ea13fd3025d5fd2bdd8b557\": rpc error: code = NotFound desc = could not find container \"cbcee76f688c973d2de72244c4863f1145ed52772ea13fd3025d5fd2bdd8b557\": container with ID starting with cbcee76f688c973d2de72244c4863f1145ed52772ea13fd3025d5fd2bdd8b557 not found: ID does not exist" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.252812 4873 scope.go:117] "RemoveContainer" containerID="5320235a23db62440b92a97af8ee2daef5081e6943bb7cb26244b3824d6d215a" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.253191 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5320235a23db62440b92a97af8ee2daef5081e6943bb7cb26244b3824d6d215a"} err="failed to get container status \"5320235a23db62440b92a97af8ee2daef5081e6943bb7cb26244b3824d6d215a\": rpc error: code = NotFound desc = could not find container \"5320235a23db62440b92a97af8ee2daef5081e6943bb7cb26244b3824d6d215a\": container with ID starting with 5320235a23db62440b92a97af8ee2daef5081e6943bb7cb26244b3824d6d215a not found: ID does not exist" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.253247 4873 scope.go:117] "RemoveContainer" containerID="c8399d5ddeaad30fb97eb909507984286dce7ccfac74ca7f7ef408cfa836976a" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.253588 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8399d5ddeaad30fb97eb909507984286dce7ccfac74ca7f7ef408cfa836976a"} err="failed to get container status \"c8399d5ddeaad30fb97eb909507984286dce7ccfac74ca7f7ef408cfa836976a\": rpc error: code = NotFound desc = could not find container \"c8399d5ddeaad30fb97eb909507984286dce7ccfac74ca7f7ef408cfa836976a\": container with ID starting with c8399d5ddeaad30fb97eb909507984286dce7ccfac74ca7f7ef408cfa836976a not found: ID does not exist" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.253611 4873 scope.go:117] "RemoveContainer" containerID="f1c8856b9b92a20b6d58ccb9e8add6d1e249ba9f5b36cef739db859b609430ed" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.253879 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1c8856b9b92a20b6d58ccb9e8add6d1e249ba9f5b36cef739db859b609430ed"} err="failed to get container status \"f1c8856b9b92a20b6d58ccb9e8add6d1e249ba9f5b36cef739db859b609430ed\": rpc error: code = NotFound desc = could not find container \"f1c8856b9b92a20b6d58ccb9e8add6d1e249ba9f5b36cef739db859b609430ed\": container with ID starting with f1c8856b9b92a20b6d58ccb9e8add6d1e249ba9f5b36cef739db859b609430ed not found: ID does not exist" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.253900 4873 scope.go:117] "RemoveContainer" containerID="cbcee76f688c973d2de72244c4863f1145ed52772ea13fd3025d5fd2bdd8b557" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.254314 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cbcee76f688c973d2de72244c4863f1145ed52772ea13fd3025d5fd2bdd8b557"} err="failed to get container status \"cbcee76f688c973d2de72244c4863f1145ed52772ea13fd3025d5fd2bdd8b557\": rpc error: code = NotFound desc = could not find container \"cbcee76f688c973d2de72244c4863f1145ed52772ea13fd3025d5fd2bdd8b557\": container with ID starting with cbcee76f688c973d2de72244c4863f1145ed52772ea13fd3025d5fd2bdd8b557 not found: ID does not exist" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.254342 4873 scope.go:117] "RemoveContainer" containerID="5320235a23db62440b92a97af8ee2daef5081e6943bb7cb26244b3824d6d215a" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.254630 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5320235a23db62440b92a97af8ee2daef5081e6943bb7cb26244b3824d6d215a"} err="failed to get container status \"5320235a23db62440b92a97af8ee2daef5081e6943bb7cb26244b3824d6d215a\": rpc error: code = NotFound desc = could not find container \"5320235a23db62440b92a97af8ee2daef5081e6943bb7cb26244b3824d6d215a\": container with ID starting with 5320235a23db62440b92a97af8ee2daef5081e6943bb7cb26244b3824d6d215a not found: ID does not exist" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.254652 4873 scope.go:117] "RemoveContainer" containerID="c8399d5ddeaad30fb97eb909507984286dce7ccfac74ca7f7ef408cfa836976a" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.255034 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8399d5ddeaad30fb97eb909507984286dce7ccfac74ca7f7ef408cfa836976a"} err="failed to get container status \"c8399d5ddeaad30fb97eb909507984286dce7ccfac74ca7f7ef408cfa836976a\": rpc error: code = NotFound desc = could not find container \"c8399d5ddeaad30fb97eb909507984286dce7ccfac74ca7f7ef408cfa836976a\": container with ID starting with c8399d5ddeaad30fb97eb909507984286dce7ccfac74ca7f7ef408cfa836976a not found: ID does not exist" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.255058 4873 scope.go:117] "RemoveContainer" containerID="f1c8856b9b92a20b6d58ccb9e8add6d1e249ba9f5b36cef739db859b609430ed" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.255614 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1c8856b9b92a20b6d58ccb9e8add6d1e249ba9f5b36cef739db859b609430ed"} err="failed to get container status \"f1c8856b9b92a20b6d58ccb9e8add6d1e249ba9f5b36cef739db859b609430ed\": rpc error: code = NotFound desc = could not find container \"f1c8856b9b92a20b6d58ccb9e8add6d1e249ba9f5b36cef739db859b609430ed\": container with ID starting with f1c8856b9b92a20b6d58ccb9e8add6d1e249ba9f5b36cef739db859b609430ed not found: ID does not exist" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.255638 4873 scope.go:117] "RemoveContainer" containerID="cbcee76f688c973d2de72244c4863f1145ed52772ea13fd3025d5fd2bdd8b557" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.255944 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cbcee76f688c973d2de72244c4863f1145ed52772ea13fd3025d5fd2bdd8b557"} err="failed to get container status \"cbcee76f688c973d2de72244c4863f1145ed52772ea13fd3025d5fd2bdd8b557\": rpc error: code = NotFound desc = could not find container \"cbcee76f688c973d2de72244c4863f1145ed52772ea13fd3025d5fd2bdd8b557\": container with ID starting with cbcee76f688c973d2de72244c4863f1145ed52772ea13fd3025d5fd2bdd8b557 not found: ID does not exist" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.255963 4873 scope.go:117] "RemoveContainer" containerID="5320235a23db62440b92a97af8ee2daef5081e6943bb7cb26244b3824d6d215a" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.256342 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5320235a23db62440b92a97af8ee2daef5081e6943bb7cb26244b3824d6d215a"} err="failed to get container status \"5320235a23db62440b92a97af8ee2daef5081e6943bb7cb26244b3824d6d215a\": rpc error: code = NotFound desc = could not find container \"5320235a23db62440b92a97af8ee2daef5081e6943bb7cb26244b3824d6d215a\": container with ID starting with 5320235a23db62440b92a97af8ee2daef5081e6943bb7cb26244b3824d6d215a not found: ID does not exist" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.256363 4873 scope.go:117] "RemoveContainer" containerID="c8399d5ddeaad30fb97eb909507984286dce7ccfac74ca7f7ef408cfa836976a" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.256671 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8399d5ddeaad30fb97eb909507984286dce7ccfac74ca7f7ef408cfa836976a"} err="failed to get container status \"c8399d5ddeaad30fb97eb909507984286dce7ccfac74ca7f7ef408cfa836976a\": rpc error: code = NotFound desc = could not find container \"c8399d5ddeaad30fb97eb909507984286dce7ccfac74ca7f7ef408cfa836976a\": container with ID starting with c8399d5ddeaad30fb97eb909507984286dce7ccfac74ca7f7ef408cfa836976a not found: ID does not exist" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.358723 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.378044 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.396490 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 09:00:54 crc kubenswrapper[4873]: E1201 09:00:54.397683 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b918b2a5-a989-44af-918c-850b1da98aa6" containerName="ceilometer-notification-agent" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.397718 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="b918b2a5-a989-44af-918c-850b1da98aa6" containerName="ceilometer-notification-agent" Dec 01 09:00:54 crc kubenswrapper[4873]: E1201 09:00:54.397798 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b918b2a5-a989-44af-918c-850b1da98aa6" containerName="proxy-httpd" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.397810 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="b918b2a5-a989-44af-918c-850b1da98aa6" containerName="proxy-httpd" Dec 01 09:00:54 crc kubenswrapper[4873]: E1201 09:00:54.397826 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b918b2a5-a989-44af-918c-850b1da98aa6" containerName="sg-core" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.397838 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="b918b2a5-a989-44af-918c-850b1da98aa6" containerName="sg-core" Dec 01 09:00:54 crc kubenswrapper[4873]: E1201 09:00:54.397866 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b918b2a5-a989-44af-918c-850b1da98aa6" containerName="ceilometer-central-agent" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.397878 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="b918b2a5-a989-44af-918c-850b1da98aa6" containerName="ceilometer-central-agent" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.401428 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="b918b2a5-a989-44af-918c-850b1da98aa6" containerName="ceilometer-central-agent" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.401522 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="b918b2a5-a989-44af-918c-850b1da98aa6" containerName="proxy-httpd" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.401552 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="b918b2a5-a989-44af-918c-850b1da98aa6" containerName="ceilometer-notification-agent" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.401577 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="b918b2a5-a989-44af-918c-850b1da98aa6" containerName="sg-core" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.407582 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.417607 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.418566 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.441294 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43cb2357-1bfa-40e6-8754-3b41c1e82034-scripts\") pod \"ceilometer-0\" (UID: \"43cb2357-1bfa-40e6-8754-3b41c1e82034\") " pod="openstack/ceilometer-0" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.441750 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/43cb2357-1bfa-40e6-8754-3b41c1e82034-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"43cb2357-1bfa-40e6-8754-3b41c1e82034\") " pod="openstack/ceilometer-0" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.441933 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43cb2357-1bfa-40e6-8754-3b41c1e82034-config-data\") pod \"ceilometer-0\" (UID: \"43cb2357-1bfa-40e6-8754-3b41c1e82034\") " pod="openstack/ceilometer-0" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.442180 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/43cb2357-1bfa-40e6-8754-3b41c1e82034-run-httpd\") pod \"ceilometer-0\" (UID: \"43cb2357-1bfa-40e6-8754-3b41c1e82034\") " pod="openstack/ceilometer-0" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.442423 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/43cb2357-1bfa-40e6-8754-3b41c1e82034-log-httpd\") pod \"ceilometer-0\" (UID: \"43cb2357-1bfa-40e6-8754-3b41c1e82034\") " pod="openstack/ceilometer-0" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.442550 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhhjx\" (UniqueName: \"kubernetes.io/projected/43cb2357-1bfa-40e6-8754-3b41c1e82034-kube-api-access-lhhjx\") pod \"ceilometer-0\" (UID: \"43cb2357-1bfa-40e6-8754-3b41c1e82034\") " pod="openstack/ceilometer-0" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.442635 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43cb2357-1bfa-40e6-8754-3b41c1e82034-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"43cb2357-1bfa-40e6-8754-3b41c1e82034\") " pod="openstack/ceilometer-0" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.449195 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b918b2a5-a989-44af-918c-850b1da98aa6" path="/var/lib/kubelet/pods/b918b2a5-a989-44af-918c-850b1da98aa6/volumes" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.450161 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.545258 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/43cb2357-1bfa-40e6-8754-3b41c1e82034-log-httpd\") pod \"ceilometer-0\" (UID: \"43cb2357-1bfa-40e6-8754-3b41c1e82034\") " pod="openstack/ceilometer-0" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.545314 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhhjx\" (UniqueName: \"kubernetes.io/projected/43cb2357-1bfa-40e6-8754-3b41c1e82034-kube-api-access-lhhjx\") pod \"ceilometer-0\" (UID: \"43cb2357-1bfa-40e6-8754-3b41c1e82034\") " pod="openstack/ceilometer-0" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.545340 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43cb2357-1bfa-40e6-8754-3b41c1e82034-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"43cb2357-1bfa-40e6-8754-3b41c1e82034\") " pod="openstack/ceilometer-0" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.545430 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43cb2357-1bfa-40e6-8754-3b41c1e82034-scripts\") pod \"ceilometer-0\" (UID: \"43cb2357-1bfa-40e6-8754-3b41c1e82034\") " pod="openstack/ceilometer-0" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.545461 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/43cb2357-1bfa-40e6-8754-3b41c1e82034-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"43cb2357-1bfa-40e6-8754-3b41c1e82034\") " pod="openstack/ceilometer-0" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.545484 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43cb2357-1bfa-40e6-8754-3b41c1e82034-config-data\") pod \"ceilometer-0\" (UID: \"43cb2357-1bfa-40e6-8754-3b41c1e82034\") " pod="openstack/ceilometer-0" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.545523 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/43cb2357-1bfa-40e6-8754-3b41c1e82034-run-httpd\") pod \"ceilometer-0\" (UID: \"43cb2357-1bfa-40e6-8754-3b41c1e82034\") " pod="openstack/ceilometer-0" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.545942 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/43cb2357-1bfa-40e6-8754-3b41c1e82034-run-httpd\") pod \"ceilometer-0\" (UID: \"43cb2357-1bfa-40e6-8754-3b41c1e82034\") " pod="openstack/ceilometer-0" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.547623 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/43cb2357-1bfa-40e6-8754-3b41c1e82034-log-httpd\") pod \"ceilometer-0\" (UID: \"43cb2357-1bfa-40e6-8754-3b41c1e82034\") " pod="openstack/ceilometer-0" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.552182 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43cb2357-1bfa-40e6-8754-3b41c1e82034-scripts\") pod \"ceilometer-0\" (UID: \"43cb2357-1bfa-40e6-8754-3b41c1e82034\") " pod="openstack/ceilometer-0" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.561170 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43cb2357-1bfa-40e6-8754-3b41c1e82034-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"43cb2357-1bfa-40e6-8754-3b41c1e82034\") " pod="openstack/ceilometer-0" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.561981 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43cb2357-1bfa-40e6-8754-3b41c1e82034-config-data\") pod \"ceilometer-0\" (UID: \"43cb2357-1bfa-40e6-8754-3b41c1e82034\") " pod="openstack/ceilometer-0" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.567323 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/43cb2357-1bfa-40e6-8754-3b41c1e82034-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"43cb2357-1bfa-40e6-8754-3b41c1e82034\") " pod="openstack/ceilometer-0" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.569722 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhhjx\" (UniqueName: \"kubernetes.io/projected/43cb2357-1bfa-40e6-8754-3b41c1e82034-kube-api-access-lhhjx\") pod \"ceilometer-0\" (UID: \"43cb2357-1bfa-40e6-8754-3b41c1e82034\") " pod="openstack/ceilometer-0" Dec 01 09:00:54 crc kubenswrapper[4873]: I1201 09:00:54.734036 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 09:00:55 crc kubenswrapper[4873]: I1201 09:00:55.241726 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 09:00:55 crc kubenswrapper[4873]: W1201 09:00:55.251307 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod43cb2357_1bfa_40e6_8754_3b41c1e82034.slice/crio-fbe57d240c40eaba3f0d58754d18ee525cdca2032064f86da5ff67f123eabc69 WatchSource:0}: Error finding container fbe57d240c40eaba3f0d58754d18ee525cdca2032064f86da5ff67f123eabc69: Status 404 returned error can't find the container with id fbe57d240c40eaba3f0d58754d18ee525cdca2032064f86da5ff67f123eabc69 Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.034942 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"43cb2357-1bfa-40e6-8754-3b41c1e82034","Type":"ContainerStarted","Data":"fbe57d240c40eaba3f0d58754d18ee525cdca2032064f86da5ff67f123eabc69"} Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.390257 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-w2kdn"] Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.392539 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-w2kdn" Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.408101 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-w2kdn"] Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.450783 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-231a-account-create-update-7vv9t"] Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.452388 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-231a-account-create-update-7vv9t" Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.459396 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.492303 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0539778e-b89c-4ae6-8a79-7cb4578a0c1c-operator-scripts\") pod \"nova-api-db-create-w2kdn\" (UID: \"0539778e-b89c-4ae6-8a79-7cb4578a0c1c\") " pod="openstack/nova-api-db-create-w2kdn" Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.492419 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/951312e8-93e4-42b5-8d28-f5df19c6a2d3-operator-scripts\") pod \"nova-api-231a-account-create-update-7vv9t\" (UID: \"951312e8-93e4-42b5-8d28-f5df19c6a2d3\") " pod="openstack/nova-api-231a-account-create-update-7vv9t" Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.492553 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p78rf\" (UniqueName: \"kubernetes.io/projected/951312e8-93e4-42b5-8d28-f5df19c6a2d3-kube-api-access-p78rf\") pod \"nova-api-231a-account-create-update-7vv9t\" (UID: \"951312e8-93e4-42b5-8d28-f5df19c6a2d3\") " pod="openstack/nova-api-231a-account-create-update-7vv9t" Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.492598 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fw9ps\" (UniqueName: \"kubernetes.io/projected/0539778e-b89c-4ae6-8a79-7cb4578a0c1c-kube-api-access-fw9ps\") pod \"nova-api-db-create-w2kdn\" (UID: \"0539778e-b89c-4ae6-8a79-7cb4578a0c1c\") " pod="openstack/nova-api-db-create-w2kdn" Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.502345 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-231a-account-create-update-7vv9t"] Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.520405 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-g6w99"] Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.522246 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-g6w99" Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.539320 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-g6w99"] Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.594281 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gjms\" (UniqueName: \"kubernetes.io/projected/a21563fd-0cda-4e31-af93-976c1950e56c-kube-api-access-5gjms\") pod \"nova-cell0-db-create-g6w99\" (UID: \"a21563fd-0cda-4e31-af93-976c1950e56c\") " pod="openstack/nova-cell0-db-create-g6w99" Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.594349 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0539778e-b89c-4ae6-8a79-7cb4578a0c1c-operator-scripts\") pod \"nova-api-db-create-w2kdn\" (UID: \"0539778e-b89c-4ae6-8a79-7cb4578a0c1c\") " pod="openstack/nova-api-db-create-w2kdn" Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.594397 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/951312e8-93e4-42b5-8d28-f5df19c6a2d3-operator-scripts\") pod \"nova-api-231a-account-create-update-7vv9t\" (UID: \"951312e8-93e4-42b5-8d28-f5df19c6a2d3\") " pod="openstack/nova-api-231a-account-create-update-7vv9t" Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.594561 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p78rf\" (UniqueName: \"kubernetes.io/projected/951312e8-93e4-42b5-8d28-f5df19c6a2d3-kube-api-access-p78rf\") pod \"nova-api-231a-account-create-update-7vv9t\" (UID: \"951312e8-93e4-42b5-8d28-f5df19c6a2d3\") " pod="openstack/nova-api-231a-account-create-update-7vv9t" Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.594631 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a21563fd-0cda-4e31-af93-976c1950e56c-operator-scripts\") pod \"nova-cell0-db-create-g6w99\" (UID: \"a21563fd-0cda-4e31-af93-976c1950e56c\") " pod="openstack/nova-cell0-db-create-g6w99" Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.594731 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fw9ps\" (UniqueName: \"kubernetes.io/projected/0539778e-b89c-4ae6-8a79-7cb4578a0c1c-kube-api-access-fw9ps\") pod \"nova-api-db-create-w2kdn\" (UID: \"0539778e-b89c-4ae6-8a79-7cb4578a0c1c\") " pod="openstack/nova-api-db-create-w2kdn" Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.595416 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0539778e-b89c-4ae6-8a79-7cb4578a0c1c-operator-scripts\") pod \"nova-api-db-create-w2kdn\" (UID: \"0539778e-b89c-4ae6-8a79-7cb4578a0c1c\") " pod="openstack/nova-api-db-create-w2kdn" Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.595454 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/951312e8-93e4-42b5-8d28-f5df19c6a2d3-operator-scripts\") pod \"nova-api-231a-account-create-update-7vv9t\" (UID: \"951312e8-93e4-42b5-8d28-f5df19c6a2d3\") " pod="openstack/nova-api-231a-account-create-update-7vv9t" Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.598308 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-01d7-account-create-update-rnjjk"] Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.599725 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-01d7-account-create-update-rnjjk" Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.603682 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.609843 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-01d7-account-create-update-rnjjk"] Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.624458 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fw9ps\" (UniqueName: \"kubernetes.io/projected/0539778e-b89c-4ae6-8a79-7cb4578a0c1c-kube-api-access-fw9ps\") pod \"nova-api-db-create-w2kdn\" (UID: \"0539778e-b89c-4ae6-8a79-7cb4578a0c1c\") " pod="openstack/nova-api-db-create-w2kdn" Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.627697 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p78rf\" (UniqueName: \"kubernetes.io/projected/951312e8-93e4-42b5-8d28-f5df19c6a2d3-kube-api-access-p78rf\") pod \"nova-api-231a-account-create-update-7vv9t\" (UID: \"951312e8-93e4-42b5-8d28-f5df19c6a2d3\") " pod="openstack/nova-api-231a-account-create-update-7vv9t" Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.696194 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a21563fd-0cda-4e31-af93-976c1950e56c-operator-scripts\") pod \"nova-cell0-db-create-g6w99\" (UID: \"a21563fd-0cda-4e31-af93-976c1950e56c\") " pod="openstack/nova-cell0-db-create-g6w99" Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.696355 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gjms\" (UniqueName: \"kubernetes.io/projected/a21563fd-0cda-4e31-af93-976c1950e56c-kube-api-access-5gjms\") pod \"nova-cell0-db-create-g6w99\" (UID: \"a21563fd-0cda-4e31-af93-976c1950e56c\") " pod="openstack/nova-cell0-db-create-g6w99" Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.697997 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a21563fd-0cda-4e31-af93-976c1950e56c-operator-scripts\") pod \"nova-cell0-db-create-g6w99\" (UID: \"a21563fd-0cda-4e31-af93-976c1950e56c\") " pod="openstack/nova-cell0-db-create-g6w99" Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.714754 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gjms\" (UniqueName: \"kubernetes.io/projected/a21563fd-0cda-4e31-af93-976c1950e56c-kube-api-access-5gjms\") pod \"nova-cell0-db-create-g6w99\" (UID: \"a21563fd-0cda-4e31-af93-976c1950e56c\") " pod="openstack/nova-cell0-db-create-g6w99" Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.717783 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-ld88d"] Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.719257 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-ld88d" Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.733906 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-2df0-account-create-update-k6xbx"] Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.735637 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-2df0-account-create-update-k6xbx" Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.740872 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.748992 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-2df0-account-create-update-k6xbx"] Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.754615 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-w2kdn" Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.758906 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-ld88d"] Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.782614 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-231a-account-create-update-7vv9t" Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.800104 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/76f16e3e-d8e2-4585-aa95-61390b75713b-operator-scripts\") pod \"nova-cell0-01d7-account-create-update-rnjjk\" (UID: \"76f16e3e-d8e2-4585-aa95-61390b75713b\") " pod="openstack/nova-cell0-01d7-account-create-update-rnjjk" Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.800731 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxgcp\" (UniqueName: \"kubernetes.io/projected/76f16e3e-d8e2-4585-aa95-61390b75713b-kube-api-access-qxgcp\") pod \"nova-cell0-01d7-account-create-update-rnjjk\" (UID: \"76f16e3e-d8e2-4585-aa95-61390b75713b\") " pod="openstack/nova-cell0-01d7-account-create-update-rnjjk" Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.846822 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-g6w99" Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.904523 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ff3eb73-9732-4b92-a21c-3a195b39f791-operator-scripts\") pod \"nova-cell1-2df0-account-create-update-k6xbx\" (UID: \"1ff3eb73-9732-4b92-a21c-3a195b39f791\") " pod="openstack/nova-cell1-2df0-account-create-update-k6xbx" Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.905175 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/76f16e3e-d8e2-4585-aa95-61390b75713b-operator-scripts\") pod \"nova-cell0-01d7-account-create-update-rnjjk\" (UID: \"76f16e3e-d8e2-4585-aa95-61390b75713b\") " pod="openstack/nova-cell0-01d7-account-create-update-rnjjk" Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.905253 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxgcp\" (UniqueName: \"kubernetes.io/projected/76f16e3e-d8e2-4585-aa95-61390b75713b-kube-api-access-qxgcp\") pod \"nova-cell0-01d7-account-create-update-rnjjk\" (UID: \"76f16e3e-d8e2-4585-aa95-61390b75713b\") " pod="openstack/nova-cell0-01d7-account-create-update-rnjjk" Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.905309 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpdws\" (UniqueName: \"kubernetes.io/projected/1ff3eb73-9732-4b92-a21c-3a195b39f791-kube-api-access-fpdws\") pod \"nova-cell1-2df0-account-create-update-k6xbx\" (UID: \"1ff3eb73-9732-4b92-a21c-3a195b39f791\") " pod="openstack/nova-cell1-2df0-account-create-update-k6xbx" Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.905457 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zchl\" (UniqueName: \"kubernetes.io/projected/d94ba6ed-6043-486e-bb35-8c14add87f61-kube-api-access-5zchl\") pod \"nova-cell1-db-create-ld88d\" (UID: \"d94ba6ed-6043-486e-bb35-8c14add87f61\") " pod="openstack/nova-cell1-db-create-ld88d" Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.905534 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d94ba6ed-6043-486e-bb35-8c14add87f61-operator-scripts\") pod \"nova-cell1-db-create-ld88d\" (UID: \"d94ba6ed-6043-486e-bb35-8c14add87f61\") " pod="openstack/nova-cell1-db-create-ld88d" Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.908536 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/76f16e3e-d8e2-4585-aa95-61390b75713b-operator-scripts\") pod \"nova-cell0-01d7-account-create-update-rnjjk\" (UID: \"76f16e3e-d8e2-4585-aa95-61390b75713b\") " pod="openstack/nova-cell0-01d7-account-create-update-rnjjk" Dec 01 09:00:56 crc kubenswrapper[4873]: I1201 09:00:56.943886 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxgcp\" (UniqueName: \"kubernetes.io/projected/76f16e3e-d8e2-4585-aa95-61390b75713b-kube-api-access-qxgcp\") pod \"nova-cell0-01d7-account-create-update-rnjjk\" (UID: \"76f16e3e-d8e2-4585-aa95-61390b75713b\") " pod="openstack/nova-cell0-01d7-account-create-update-rnjjk" Dec 01 09:00:57 crc kubenswrapper[4873]: I1201 09:00:57.008587 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zchl\" (UniqueName: \"kubernetes.io/projected/d94ba6ed-6043-486e-bb35-8c14add87f61-kube-api-access-5zchl\") pod \"nova-cell1-db-create-ld88d\" (UID: \"d94ba6ed-6043-486e-bb35-8c14add87f61\") " pod="openstack/nova-cell1-db-create-ld88d" Dec 01 09:00:57 crc kubenswrapper[4873]: I1201 09:00:57.008664 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d94ba6ed-6043-486e-bb35-8c14add87f61-operator-scripts\") pod \"nova-cell1-db-create-ld88d\" (UID: \"d94ba6ed-6043-486e-bb35-8c14add87f61\") " pod="openstack/nova-cell1-db-create-ld88d" Dec 01 09:00:57 crc kubenswrapper[4873]: I1201 09:00:57.008725 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ff3eb73-9732-4b92-a21c-3a195b39f791-operator-scripts\") pod \"nova-cell1-2df0-account-create-update-k6xbx\" (UID: \"1ff3eb73-9732-4b92-a21c-3a195b39f791\") " pod="openstack/nova-cell1-2df0-account-create-update-k6xbx" Dec 01 09:00:57 crc kubenswrapper[4873]: I1201 09:00:57.008786 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpdws\" (UniqueName: \"kubernetes.io/projected/1ff3eb73-9732-4b92-a21c-3a195b39f791-kube-api-access-fpdws\") pod \"nova-cell1-2df0-account-create-update-k6xbx\" (UID: \"1ff3eb73-9732-4b92-a21c-3a195b39f791\") " pod="openstack/nova-cell1-2df0-account-create-update-k6xbx" Dec 01 09:00:57 crc kubenswrapper[4873]: I1201 09:00:57.009628 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d94ba6ed-6043-486e-bb35-8c14add87f61-operator-scripts\") pod \"nova-cell1-db-create-ld88d\" (UID: \"d94ba6ed-6043-486e-bb35-8c14add87f61\") " pod="openstack/nova-cell1-db-create-ld88d" Dec 01 09:00:57 crc kubenswrapper[4873]: I1201 09:00:57.009801 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ff3eb73-9732-4b92-a21c-3a195b39f791-operator-scripts\") pod \"nova-cell1-2df0-account-create-update-k6xbx\" (UID: \"1ff3eb73-9732-4b92-a21c-3a195b39f791\") " pod="openstack/nova-cell1-2df0-account-create-update-k6xbx" Dec 01 09:00:57 crc kubenswrapper[4873]: I1201 09:00:57.036530 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpdws\" (UniqueName: \"kubernetes.io/projected/1ff3eb73-9732-4b92-a21c-3a195b39f791-kube-api-access-fpdws\") pod \"nova-cell1-2df0-account-create-update-k6xbx\" (UID: \"1ff3eb73-9732-4b92-a21c-3a195b39f791\") " pod="openstack/nova-cell1-2df0-account-create-update-k6xbx" Dec 01 09:00:57 crc kubenswrapper[4873]: I1201 09:00:57.036894 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zchl\" (UniqueName: \"kubernetes.io/projected/d94ba6ed-6043-486e-bb35-8c14add87f61-kube-api-access-5zchl\") pod \"nova-cell1-db-create-ld88d\" (UID: \"d94ba6ed-6043-486e-bb35-8c14add87f61\") " pod="openstack/nova-cell1-db-create-ld88d" Dec 01 09:00:57 crc kubenswrapper[4873]: I1201 09:00:57.051174 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"43cb2357-1bfa-40e6-8754-3b41c1e82034","Type":"ContainerStarted","Data":"7517972ce6cc58991973ed9d7a63f8c3d422dd492a3c0b693ce83d9ca331e879"} Dec 01 09:00:57 crc kubenswrapper[4873]: I1201 09:00:57.069989 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-01d7-account-create-update-rnjjk" Dec 01 09:00:57 crc kubenswrapper[4873]: I1201 09:00:57.086986 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-ld88d" Dec 01 09:00:57 crc kubenswrapper[4873]: I1201 09:00:57.096706 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-2df0-account-create-update-k6xbx" Dec 01 09:00:57 crc kubenswrapper[4873]: I1201 09:00:57.251342 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-g6w99"] Dec 01 09:00:57 crc kubenswrapper[4873]: I1201 09:00:57.259061 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-w2kdn"] Dec 01 09:00:57 crc kubenswrapper[4873]: W1201 09:00:57.265289 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0539778e_b89c_4ae6_8a79_7cb4578a0c1c.slice/crio-6d1ae8b435b433c217acee8b24f02fa5e6a7a680477877f12748d1484525ca96 WatchSource:0}: Error finding container 6d1ae8b435b433c217acee8b24f02fa5e6a7a680477877f12748d1484525ca96: Status 404 returned error can't find the container with id 6d1ae8b435b433c217acee8b24f02fa5e6a7a680477877f12748d1484525ca96 Dec 01 09:00:57 crc kubenswrapper[4873]: W1201 09:00:57.267787 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda21563fd_0cda_4e31_af93_976c1950e56c.slice/crio-1f6c843b56a8674e66941ad4ec87865e4244894a21d97989a99582c75d0a5f38 WatchSource:0}: Error finding container 1f6c843b56a8674e66941ad4ec87865e4244894a21d97989a99582c75d0a5f38: Status 404 returned error can't find the container with id 1f6c843b56a8674e66941ad4ec87865e4244894a21d97989a99582c75d0a5f38 Dec 01 09:00:57 crc kubenswrapper[4873]: I1201 09:00:57.360375 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-231a-account-create-update-7vv9t"] Dec 01 09:00:57 crc kubenswrapper[4873]: I1201 09:00:57.563072 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-2df0-account-create-update-k6xbx"] Dec 01 09:00:57 crc kubenswrapper[4873]: W1201 09:00:57.588844 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1ff3eb73_9732_4b92_a21c_3a195b39f791.slice/crio-ce0e74560c7fb436295a3f8e6d83ee5fe2ffa8067a5c8691714a400b5a32154b WatchSource:0}: Error finding container ce0e74560c7fb436295a3f8e6d83ee5fe2ffa8067a5c8691714a400b5a32154b: Status 404 returned error can't find the container with id ce0e74560c7fb436295a3f8e6d83ee5fe2ffa8067a5c8691714a400b5a32154b Dec 01 09:00:57 crc kubenswrapper[4873]: I1201 09:00:57.873882 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-01d7-account-create-update-rnjjk"] Dec 01 09:00:58 crc kubenswrapper[4873]: I1201 09:00:58.068806 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-ld88d"] Dec 01 09:00:58 crc kubenswrapper[4873]: I1201 09:00:58.071514 4873 generic.go:334] "Generic (PLEG): container finished" podID="a21563fd-0cda-4e31-af93-976c1950e56c" containerID="6d96f4b4c03294ca19a5cac036c6ad6e9bdc9fbff00ac03ebb291a0f6853705f" exitCode=0 Dec 01 09:00:58 crc kubenswrapper[4873]: I1201 09:00:58.071582 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-g6w99" event={"ID":"a21563fd-0cda-4e31-af93-976c1950e56c","Type":"ContainerDied","Data":"6d96f4b4c03294ca19a5cac036c6ad6e9bdc9fbff00ac03ebb291a0f6853705f"} Dec 01 09:00:58 crc kubenswrapper[4873]: I1201 09:00:58.071632 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-g6w99" event={"ID":"a21563fd-0cda-4e31-af93-976c1950e56c","Type":"ContainerStarted","Data":"1f6c843b56a8674e66941ad4ec87865e4244894a21d97989a99582c75d0a5f38"} Dec 01 09:00:58 crc kubenswrapper[4873]: I1201 09:00:58.080775 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-2df0-account-create-update-k6xbx" event={"ID":"1ff3eb73-9732-4b92-a21c-3a195b39f791","Type":"ContainerStarted","Data":"29199103bf26d5451c7186e11bf49db090317374b7bcc88aceb00d4ec66e563e"} Dec 01 09:00:58 crc kubenswrapper[4873]: I1201 09:00:58.080825 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-2df0-account-create-update-k6xbx" event={"ID":"1ff3eb73-9732-4b92-a21c-3a195b39f791","Type":"ContainerStarted","Data":"ce0e74560c7fb436295a3f8e6d83ee5fe2ffa8067a5c8691714a400b5a32154b"} Dec 01 09:00:58 crc kubenswrapper[4873]: I1201 09:00:58.082859 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"43cb2357-1bfa-40e6-8754-3b41c1e82034","Type":"ContainerStarted","Data":"eb3e66a154eef8f44e0931a6a8c623db26f6c343b97e67f475825e9a46c8632a"} Dec 01 09:00:58 crc kubenswrapper[4873]: I1201 09:00:58.084044 4873 generic.go:334] "Generic (PLEG): container finished" podID="0539778e-b89c-4ae6-8a79-7cb4578a0c1c" containerID="523238858a5e7d371b2e3b8d87e743130e7c243cfc9d9ebddaf8d98fb42483a2" exitCode=0 Dec 01 09:00:58 crc kubenswrapper[4873]: I1201 09:00:58.084092 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-w2kdn" event={"ID":"0539778e-b89c-4ae6-8a79-7cb4578a0c1c","Type":"ContainerDied","Data":"523238858a5e7d371b2e3b8d87e743130e7c243cfc9d9ebddaf8d98fb42483a2"} Dec 01 09:00:58 crc kubenswrapper[4873]: I1201 09:00:58.084111 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-w2kdn" event={"ID":"0539778e-b89c-4ae6-8a79-7cb4578a0c1c","Type":"ContainerStarted","Data":"6d1ae8b435b433c217acee8b24f02fa5e6a7a680477877f12748d1484525ca96"} Dec 01 09:00:58 crc kubenswrapper[4873]: I1201 09:00:58.085367 4873 generic.go:334] "Generic (PLEG): container finished" podID="951312e8-93e4-42b5-8d28-f5df19c6a2d3" containerID="928b07338f76deaadc48dcb860deff8b6ee9693b1b0dd0ab0d01c63540831087" exitCode=0 Dec 01 09:00:58 crc kubenswrapper[4873]: I1201 09:00:58.085408 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-231a-account-create-update-7vv9t" event={"ID":"951312e8-93e4-42b5-8d28-f5df19c6a2d3","Type":"ContainerDied","Data":"928b07338f76deaadc48dcb860deff8b6ee9693b1b0dd0ab0d01c63540831087"} Dec 01 09:00:58 crc kubenswrapper[4873]: I1201 09:00:58.085425 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-231a-account-create-update-7vv9t" event={"ID":"951312e8-93e4-42b5-8d28-f5df19c6a2d3","Type":"ContainerStarted","Data":"826cdb0c048ceea466e43504d3dc05ce076316c3abdcfa16f420bd344d5dcc5c"} Dec 01 09:00:58 crc kubenswrapper[4873]: I1201 09:00:58.087934 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-01d7-account-create-update-rnjjk" event={"ID":"76f16e3e-d8e2-4585-aa95-61390b75713b","Type":"ContainerStarted","Data":"e61896b9e53e1a39bc0ef163c5b70a472241364dbb3fad2d4672794113cc7319"} Dec 01 09:00:58 crc kubenswrapper[4873]: W1201 09:00:58.116495 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd94ba6ed_6043_486e_bb35_8c14add87f61.slice/crio-5377def4e9869af146c36a76cf8d5fdd4cdea82b2c96c2d112326a1fb441f406 WatchSource:0}: Error finding container 5377def4e9869af146c36a76cf8d5fdd4cdea82b2c96c2d112326a1fb441f406: Status 404 returned error can't find the container with id 5377def4e9869af146c36a76cf8d5fdd4cdea82b2c96c2d112326a1fb441f406 Dec 01 09:00:58 crc kubenswrapper[4873]: I1201 09:00:58.138459 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-2df0-account-create-update-k6xbx" podStartSLOduration=2.138422764 podStartE2EDuration="2.138422764s" podCreationTimestamp="2025-12-01 09:00:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:00:58.131362391 +0000 UTC m=+1234.033470950" watchObservedRunningTime="2025-12-01 09:00:58.138422764 +0000 UTC m=+1234.040531303" Dec 01 09:00:59 crc kubenswrapper[4873]: I1201 09:00:59.101202 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"43cb2357-1bfa-40e6-8754-3b41c1e82034","Type":"ContainerStarted","Data":"972ebb4605647c3a76cc89e424ba37b80cb0950be55338668b5f488e08f760a3"} Dec 01 09:00:59 crc kubenswrapper[4873]: I1201 09:00:59.103770 4873 generic.go:334] "Generic (PLEG): container finished" podID="76f16e3e-d8e2-4585-aa95-61390b75713b" containerID="5ba249e9ca335fc03a2d70f8e3a190f77197791d48ef944e6c4d4574c5c5276e" exitCode=0 Dec 01 09:00:59 crc kubenswrapper[4873]: I1201 09:00:59.103862 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-01d7-account-create-update-rnjjk" event={"ID":"76f16e3e-d8e2-4585-aa95-61390b75713b","Type":"ContainerDied","Data":"5ba249e9ca335fc03a2d70f8e3a190f77197791d48ef944e6c4d4574c5c5276e"} Dec 01 09:00:59 crc kubenswrapper[4873]: I1201 09:00:59.106541 4873 generic.go:334] "Generic (PLEG): container finished" podID="d94ba6ed-6043-486e-bb35-8c14add87f61" containerID="cbc26a4e9818a65332ccad8daf87f3c64050554ff7c97689265a4d9c5971f653" exitCode=0 Dec 01 09:00:59 crc kubenswrapper[4873]: I1201 09:00:59.106619 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-ld88d" event={"ID":"d94ba6ed-6043-486e-bb35-8c14add87f61","Type":"ContainerDied","Data":"cbc26a4e9818a65332ccad8daf87f3c64050554ff7c97689265a4d9c5971f653"} Dec 01 09:00:59 crc kubenswrapper[4873]: I1201 09:00:59.106649 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-ld88d" event={"ID":"d94ba6ed-6043-486e-bb35-8c14add87f61","Type":"ContainerStarted","Data":"5377def4e9869af146c36a76cf8d5fdd4cdea82b2c96c2d112326a1fb441f406"} Dec 01 09:00:59 crc kubenswrapper[4873]: I1201 09:00:59.108703 4873 generic.go:334] "Generic (PLEG): container finished" podID="1ff3eb73-9732-4b92-a21c-3a195b39f791" containerID="29199103bf26d5451c7186e11bf49db090317374b7bcc88aceb00d4ec66e563e" exitCode=0 Dec 01 09:00:59 crc kubenswrapper[4873]: I1201 09:00:59.108786 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-2df0-account-create-update-k6xbx" event={"ID":"1ff3eb73-9732-4b92-a21c-3a195b39f791","Type":"ContainerDied","Data":"29199103bf26d5451c7186e11bf49db090317374b7bcc88aceb00d4ec66e563e"} Dec 01 09:00:59 crc kubenswrapper[4873]: I1201 09:00:59.646832 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-w2kdn" Dec 01 09:00:59 crc kubenswrapper[4873]: I1201 09:00:59.675849 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fw9ps\" (UniqueName: \"kubernetes.io/projected/0539778e-b89c-4ae6-8a79-7cb4578a0c1c-kube-api-access-fw9ps\") pod \"0539778e-b89c-4ae6-8a79-7cb4578a0c1c\" (UID: \"0539778e-b89c-4ae6-8a79-7cb4578a0c1c\") " Dec 01 09:00:59 crc kubenswrapper[4873]: I1201 09:00:59.678409 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0539778e-b89c-4ae6-8a79-7cb4578a0c1c-operator-scripts\") pod \"0539778e-b89c-4ae6-8a79-7cb4578a0c1c\" (UID: \"0539778e-b89c-4ae6-8a79-7cb4578a0c1c\") " Dec 01 09:00:59 crc kubenswrapper[4873]: I1201 09:00:59.679440 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0539778e-b89c-4ae6-8a79-7cb4578a0c1c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0539778e-b89c-4ae6-8a79-7cb4578a0c1c" (UID: "0539778e-b89c-4ae6-8a79-7cb4578a0c1c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:00:59 crc kubenswrapper[4873]: I1201 09:00:59.688746 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0539778e-b89c-4ae6-8a79-7cb4578a0c1c-kube-api-access-fw9ps" (OuterVolumeSpecName: "kube-api-access-fw9ps") pod "0539778e-b89c-4ae6-8a79-7cb4578a0c1c" (UID: "0539778e-b89c-4ae6-8a79-7cb4578a0c1c"). InnerVolumeSpecName "kube-api-access-fw9ps". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:00:59 crc kubenswrapper[4873]: I1201 09:00:59.759561 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-g6w99" Dec 01 09:00:59 crc kubenswrapper[4873]: I1201 09:00:59.766096 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-231a-account-create-update-7vv9t" Dec 01 09:00:59 crc kubenswrapper[4873]: I1201 09:00:59.786545 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fw9ps\" (UniqueName: \"kubernetes.io/projected/0539778e-b89c-4ae6-8a79-7cb4578a0c1c-kube-api-access-fw9ps\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:59 crc kubenswrapper[4873]: I1201 09:00:59.786588 4873 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0539778e-b89c-4ae6-8a79-7cb4578a0c1c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:59 crc kubenswrapper[4873]: I1201 09:00:59.887680 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a21563fd-0cda-4e31-af93-976c1950e56c-operator-scripts\") pod \"a21563fd-0cda-4e31-af93-976c1950e56c\" (UID: \"a21563fd-0cda-4e31-af93-976c1950e56c\") " Dec 01 09:00:59 crc kubenswrapper[4873]: I1201 09:00:59.887871 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5gjms\" (UniqueName: \"kubernetes.io/projected/a21563fd-0cda-4e31-af93-976c1950e56c-kube-api-access-5gjms\") pod \"a21563fd-0cda-4e31-af93-976c1950e56c\" (UID: \"a21563fd-0cda-4e31-af93-976c1950e56c\") " Dec 01 09:00:59 crc kubenswrapper[4873]: I1201 09:00:59.887946 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p78rf\" (UniqueName: \"kubernetes.io/projected/951312e8-93e4-42b5-8d28-f5df19c6a2d3-kube-api-access-p78rf\") pod \"951312e8-93e4-42b5-8d28-f5df19c6a2d3\" (UID: \"951312e8-93e4-42b5-8d28-f5df19c6a2d3\") " Dec 01 09:00:59 crc kubenswrapper[4873]: I1201 09:00:59.888110 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/951312e8-93e4-42b5-8d28-f5df19c6a2d3-operator-scripts\") pod \"951312e8-93e4-42b5-8d28-f5df19c6a2d3\" (UID: \"951312e8-93e4-42b5-8d28-f5df19c6a2d3\") " Dec 01 09:00:59 crc kubenswrapper[4873]: I1201 09:00:59.889373 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/951312e8-93e4-42b5-8d28-f5df19c6a2d3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "951312e8-93e4-42b5-8d28-f5df19c6a2d3" (UID: "951312e8-93e4-42b5-8d28-f5df19c6a2d3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:00:59 crc kubenswrapper[4873]: I1201 09:00:59.889396 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a21563fd-0cda-4e31-af93-976c1950e56c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a21563fd-0cda-4e31-af93-976c1950e56c" (UID: "a21563fd-0cda-4e31-af93-976c1950e56c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:00:59 crc kubenswrapper[4873]: I1201 09:00:59.893397 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a21563fd-0cda-4e31-af93-976c1950e56c-kube-api-access-5gjms" (OuterVolumeSpecName: "kube-api-access-5gjms") pod "a21563fd-0cda-4e31-af93-976c1950e56c" (UID: "a21563fd-0cda-4e31-af93-976c1950e56c"). InnerVolumeSpecName "kube-api-access-5gjms". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:00:59 crc kubenswrapper[4873]: I1201 09:00:59.893952 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/951312e8-93e4-42b5-8d28-f5df19c6a2d3-kube-api-access-p78rf" (OuterVolumeSpecName: "kube-api-access-p78rf") pod "951312e8-93e4-42b5-8d28-f5df19c6a2d3" (UID: "951312e8-93e4-42b5-8d28-f5df19c6a2d3"). InnerVolumeSpecName "kube-api-access-p78rf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:00:59 crc kubenswrapper[4873]: I1201 09:00:59.991998 4873 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/951312e8-93e4-42b5-8d28-f5df19c6a2d3-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:59 crc kubenswrapper[4873]: I1201 09:00:59.992137 4873 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a21563fd-0cda-4e31-af93-976c1950e56c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:59 crc kubenswrapper[4873]: I1201 09:00:59.992158 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5gjms\" (UniqueName: \"kubernetes.io/projected/a21563fd-0cda-4e31-af93-976c1950e56c-kube-api-access-5gjms\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:59 crc kubenswrapper[4873]: I1201 09:00:59.992180 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p78rf\" (UniqueName: \"kubernetes.io/projected/951312e8-93e4-42b5-8d28-f5df19c6a2d3-kube-api-access-p78rf\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:00 crc kubenswrapper[4873]: I1201 09:01:00.121634 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-w2kdn" event={"ID":"0539778e-b89c-4ae6-8a79-7cb4578a0c1c","Type":"ContainerDied","Data":"6d1ae8b435b433c217acee8b24f02fa5e6a7a680477877f12748d1484525ca96"} Dec 01 09:01:00 crc kubenswrapper[4873]: I1201 09:01:00.121704 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d1ae8b435b433c217acee8b24f02fa5e6a7a680477877f12748d1484525ca96" Dec 01 09:01:00 crc kubenswrapper[4873]: I1201 09:01:00.121676 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-w2kdn" Dec 01 09:01:00 crc kubenswrapper[4873]: I1201 09:01:00.124523 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-231a-account-create-update-7vv9t" Dec 01 09:01:00 crc kubenswrapper[4873]: I1201 09:01:00.124505 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-231a-account-create-update-7vv9t" event={"ID":"951312e8-93e4-42b5-8d28-f5df19c6a2d3","Type":"ContainerDied","Data":"826cdb0c048ceea466e43504d3dc05ce076316c3abdcfa16f420bd344d5dcc5c"} Dec 01 09:01:00 crc kubenswrapper[4873]: I1201 09:01:00.124779 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="826cdb0c048ceea466e43504d3dc05ce076316c3abdcfa16f420bd344d5dcc5c" Dec 01 09:01:00 crc kubenswrapper[4873]: I1201 09:01:00.126421 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-g6w99" event={"ID":"a21563fd-0cda-4e31-af93-976c1950e56c","Type":"ContainerDied","Data":"1f6c843b56a8674e66941ad4ec87865e4244894a21d97989a99582c75d0a5f38"} Dec 01 09:01:00 crc kubenswrapper[4873]: I1201 09:01:00.126495 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1f6c843b56a8674e66941ad4ec87865e4244894a21d97989a99582c75d0a5f38" Dec 01 09:01:00 crc kubenswrapper[4873]: I1201 09:01:00.126587 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-g6w99" Dec 01 09:01:00 crc kubenswrapper[4873]: I1201 09:01:00.182497 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29409661-gls8t"] Dec 01 09:01:00 crc kubenswrapper[4873]: E1201 09:01:00.183121 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0539778e-b89c-4ae6-8a79-7cb4578a0c1c" containerName="mariadb-database-create" Dec 01 09:01:00 crc kubenswrapper[4873]: I1201 09:01:00.183144 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="0539778e-b89c-4ae6-8a79-7cb4578a0c1c" containerName="mariadb-database-create" Dec 01 09:01:00 crc kubenswrapper[4873]: E1201 09:01:00.183175 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="951312e8-93e4-42b5-8d28-f5df19c6a2d3" containerName="mariadb-account-create-update" Dec 01 09:01:00 crc kubenswrapper[4873]: I1201 09:01:00.183182 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="951312e8-93e4-42b5-8d28-f5df19c6a2d3" containerName="mariadb-account-create-update" Dec 01 09:01:00 crc kubenswrapper[4873]: E1201 09:01:00.183197 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a21563fd-0cda-4e31-af93-976c1950e56c" containerName="mariadb-database-create" Dec 01 09:01:00 crc kubenswrapper[4873]: I1201 09:01:00.183204 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="a21563fd-0cda-4e31-af93-976c1950e56c" containerName="mariadb-database-create" Dec 01 09:01:00 crc kubenswrapper[4873]: I1201 09:01:00.183432 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="951312e8-93e4-42b5-8d28-f5df19c6a2d3" containerName="mariadb-account-create-update" Dec 01 09:01:00 crc kubenswrapper[4873]: I1201 09:01:00.183458 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="0539778e-b89c-4ae6-8a79-7cb4578a0c1c" containerName="mariadb-database-create" Dec 01 09:01:00 crc kubenswrapper[4873]: I1201 09:01:00.183469 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="a21563fd-0cda-4e31-af93-976c1950e56c" containerName="mariadb-database-create" Dec 01 09:01:00 crc kubenswrapper[4873]: I1201 09:01:00.184336 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29409661-gls8t" Dec 01 09:01:00 crc kubenswrapper[4873]: I1201 09:01:00.195027 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29409661-gls8t"] Dec 01 09:01:00 crc kubenswrapper[4873]: I1201 09:01:00.297788 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf6891b3-2f06-4218-b321-5d7fac6edb7c-combined-ca-bundle\") pod \"keystone-cron-29409661-gls8t\" (UID: \"cf6891b3-2f06-4218-b321-5d7fac6edb7c\") " pod="openstack/keystone-cron-29409661-gls8t" Dec 01 09:01:00 crc kubenswrapper[4873]: I1201 09:01:00.298289 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf6891b3-2f06-4218-b321-5d7fac6edb7c-config-data\") pod \"keystone-cron-29409661-gls8t\" (UID: \"cf6891b3-2f06-4218-b321-5d7fac6edb7c\") " pod="openstack/keystone-cron-29409661-gls8t" Dec 01 09:01:00 crc kubenswrapper[4873]: I1201 09:01:00.298374 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cf6891b3-2f06-4218-b321-5d7fac6edb7c-fernet-keys\") pod \"keystone-cron-29409661-gls8t\" (UID: \"cf6891b3-2f06-4218-b321-5d7fac6edb7c\") " pod="openstack/keystone-cron-29409661-gls8t" Dec 01 09:01:00 crc kubenswrapper[4873]: I1201 09:01:00.298455 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q885j\" (UniqueName: \"kubernetes.io/projected/cf6891b3-2f06-4218-b321-5d7fac6edb7c-kube-api-access-q885j\") pod \"keystone-cron-29409661-gls8t\" (UID: \"cf6891b3-2f06-4218-b321-5d7fac6edb7c\") " pod="openstack/keystone-cron-29409661-gls8t" Dec 01 09:01:00 crc kubenswrapper[4873]: I1201 09:01:00.401218 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf6891b3-2f06-4218-b321-5d7fac6edb7c-config-data\") pod \"keystone-cron-29409661-gls8t\" (UID: \"cf6891b3-2f06-4218-b321-5d7fac6edb7c\") " pod="openstack/keystone-cron-29409661-gls8t" Dec 01 09:01:00 crc kubenswrapper[4873]: I1201 09:01:00.401285 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cf6891b3-2f06-4218-b321-5d7fac6edb7c-fernet-keys\") pod \"keystone-cron-29409661-gls8t\" (UID: \"cf6891b3-2f06-4218-b321-5d7fac6edb7c\") " pod="openstack/keystone-cron-29409661-gls8t" Dec 01 09:01:00 crc kubenswrapper[4873]: I1201 09:01:00.401342 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q885j\" (UniqueName: \"kubernetes.io/projected/cf6891b3-2f06-4218-b321-5d7fac6edb7c-kube-api-access-q885j\") pod \"keystone-cron-29409661-gls8t\" (UID: \"cf6891b3-2f06-4218-b321-5d7fac6edb7c\") " pod="openstack/keystone-cron-29409661-gls8t" Dec 01 09:01:00 crc kubenswrapper[4873]: I1201 09:01:00.401401 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf6891b3-2f06-4218-b321-5d7fac6edb7c-combined-ca-bundle\") pod \"keystone-cron-29409661-gls8t\" (UID: \"cf6891b3-2f06-4218-b321-5d7fac6edb7c\") " pod="openstack/keystone-cron-29409661-gls8t" Dec 01 09:01:00 crc kubenswrapper[4873]: I1201 09:01:00.409324 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cf6891b3-2f06-4218-b321-5d7fac6edb7c-fernet-keys\") pod \"keystone-cron-29409661-gls8t\" (UID: \"cf6891b3-2f06-4218-b321-5d7fac6edb7c\") " pod="openstack/keystone-cron-29409661-gls8t" Dec 01 09:01:00 crc kubenswrapper[4873]: I1201 09:01:00.410483 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf6891b3-2f06-4218-b321-5d7fac6edb7c-combined-ca-bundle\") pod \"keystone-cron-29409661-gls8t\" (UID: \"cf6891b3-2f06-4218-b321-5d7fac6edb7c\") " pod="openstack/keystone-cron-29409661-gls8t" Dec 01 09:01:00 crc kubenswrapper[4873]: I1201 09:01:00.414465 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf6891b3-2f06-4218-b321-5d7fac6edb7c-config-data\") pod \"keystone-cron-29409661-gls8t\" (UID: \"cf6891b3-2f06-4218-b321-5d7fac6edb7c\") " pod="openstack/keystone-cron-29409661-gls8t" Dec 01 09:01:00 crc kubenswrapper[4873]: I1201 09:01:00.425404 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q885j\" (UniqueName: \"kubernetes.io/projected/cf6891b3-2f06-4218-b321-5d7fac6edb7c-kube-api-access-q885j\") pod \"keystone-cron-29409661-gls8t\" (UID: \"cf6891b3-2f06-4218-b321-5d7fac6edb7c\") " pod="openstack/keystone-cron-29409661-gls8t" Dec 01 09:01:00 crc kubenswrapper[4873]: I1201 09:01:00.496267 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-ld88d" Dec 01 09:01:00 crc kubenswrapper[4873]: I1201 09:01:00.550665 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29409661-gls8t" Dec 01 09:01:00 crc kubenswrapper[4873]: I1201 09:01:00.605494 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zchl\" (UniqueName: \"kubernetes.io/projected/d94ba6ed-6043-486e-bb35-8c14add87f61-kube-api-access-5zchl\") pod \"d94ba6ed-6043-486e-bb35-8c14add87f61\" (UID: \"d94ba6ed-6043-486e-bb35-8c14add87f61\") " Dec 01 09:01:00 crc kubenswrapper[4873]: I1201 09:01:00.605707 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 09:01:00 crc kubenswrapper[4873]: I1201 09:01:00.605844 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d94ba6ed-6043-486e-bb35-8c14add87f61-operator-scripts\") pod \"d94ba6ed-6043-486e-bb35-8c14add87f61\" (UID: \"d94ba6ed-6043-486e-bb35-8c14add87f61\") " Dec 01 09:01:00 crc kubenswrapper[4873]: I1201 09:01:00.608461 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d94ba6ed-6043-486e-bb35-8c14add87f61-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d94ba6ed-6043-486e-bb35-8c14add87f61" (UID: "d94ba6ed-6043-486e-bb35-8c14add87f61"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:01:00 crc kubenswrapper[4873]: I1201 09:01:00.612324 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d94ba6ed-6043-486e-bb35-8c14add87f61-kube-api-access-5zchl" (OuterVolumeSpecName: "kube-api-access-5zchl") pod "d94ba6ed-6043-486e-bb35-8c14add87f61" (UID: "d94ba6ed-6043-486e-bb35-8c14add87f61"). InnerVolumeSpecName "kube-api-access-5zchl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:01:00 crc kubenswrapper[4873]: I1201 09:01:00.708684 4873 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d94ba6ed-6043-486e-bb35-8c14add87f61-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:00 crc kubenswrapper[4873]: I1201 09:01:00.708732 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zchl\" (UniqueName: \"kubernetes.io/projected/d94ba6ed-6043-486e-bb35-8c14add87f61-kube-api-access-5zchl\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:00 crc kubenswrapper[4873]: I1201 09:01:00.846973 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-2df0-account-create-update-k6xbx" Dec 01 09:01:00 crc kubenswrapper[4873]: I1201 09:01:00.856494 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-01d7-account-create-update-rnjjk" Dec 01 09:01:01 crc kubenswrapper[4873]: I1201 09:01:01.018701 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxgcp\" (UniqueName: \"kubernetes.io/projected/76f16e3e-d8e2-4585-aa95-61390b75713b-kube-api-access-qxgcp\") pod \"76f16e3e-d8e2-4585-aa95-61390b75713b\" (UID: \"76f16e3e-d8e2-4585-aa95-61390b75713b\") " Dec 01 09:01:01 crc kubenswrapper[4873]: I1201 09:01:01.018904 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ff3eb73-9732-4b92-a21c-3a195b39f791-operator-scripts\") pod \"1ff3eb73-9732-4b92-a21c-3a195b39f791\" (UID: \"1ff3eb73-9732-4b92-a21c-3a195b39f791\") " Dec 01 09:01:01 crc kubenswrapper[4873]: I1201 09:01:01.018976 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fpdws\" (UniqueName: \"kubernetes.io/projected/1ff3eb73-9732-4b92-a21c-3a195b39f791-kube-api-access-fpdws\") pod \"1ff3eb73-9732-4b92-a21c-3a195b39f791\" (UID: \"1ff3eb73-9732-4b92-a21c-3a195b39f791\") " Dec 01 09:01:01 crc kubenswrapper[4873]: I1201 09:01:01.019041 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/76f16e3e-d8e2-4585-aa95-61390b75713b-operator-scripts\") pod \"76f16e3e-d8e2-4585-aa95-61390b75713b\" (UID: \"76f16e3e-d8e2-4585-aa95-61390b75713b\") " Dec 01 09:01:01 crc kubenswrapper[4873]: I1201 09:01:01.020225 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ff3eb73-9732-4b92-a21c-3a195b39f791-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1ff3eb73-9732-4b92-a21c-3a195b39f791" (UID: "1ff3eb73-9732-4b92-a21c-3a195b39f791"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:01:01 crc kubenswrapper[4873]: I1201 09:01:01.020701 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76f16e3e-d8e2-4585-aa95-61390b75713b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "76f16e3e-d8e2-4585-aa95-61390b75713b" (UID: "76f16e3e-d8e2-4585-aa95-61390b75713b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:01:01 crc kubenswrapper[4873]: I1201 09:01:01.025730 4873 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ff3eb73-9732-4b92-a21c-3a195b39f791-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:01 crc kubenswrapper[4873]: I1201 09:01:01.025814 4873 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/76f16e3e-d8e2-4585-aa95-61390b75713b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:01 crc kubenswrapper[4873]: I1201 09:01:01.028319 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76f16e3e-d8e2-4585-aa95-61390b75713b-kube-api-access-qxgcp" (OuterVolumeSpecName: "kube-api-access-qxgcp") pod "76f16e3e-d8e2-4585-aa95-61390b75713b" (UID: "76f16e3e-d8e2-4585-aa95-61390b75713b"). InnerVolumeSpecName "kube-api-access-qxgcp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:01:01 crc kubenswrapper[4873]: I1201 09:01:01.028563 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ff3eb73-9732-4b92-a21c-3a195b39f791-kube-api-access-fpdws" (OuterVolumeSpecName: "kube-api-access-fpdws") pod "1ff3eb73-9732-4b92-a21c-3a195b39f791" (UID: "1ff3eb73-9732-4b92-a21c-3a195b39f791"). InnerVolumeSpecName "kube-api-access-fpdws". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:01:01 crc kubenswrapper[4873]: I1201 09:01:01.127750 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qxgcp\" (UniqueName: \"kubernetes.io/projected/76f16e3e-d8e2-4585-aa95-61390b75713b-kube-api-access-qxgcp\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:01 crc kubenswrapper[4873]: I1201 09:01:01.127782 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fpdws\" (UniqueName: \"kubernetes.io/projected/1ff3eb73-9732-4b92-a21c-3a195b39f791-kube-api-access-fpdws\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:01 crc kubenswrapper[4873]: I1201 09:01:01.131285 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29409661-gls8t"] Dec 01 09:01:01 crc kubenswrapper[4873]: W1201 09:01:01.137690 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcf6891b3_2f06_4218_b321_5d7fac6edb7c.slice/crio-0275f382fc99b37c5ce61f42a7205dce3fdc2c4efe5fefcf7d29fa8a8c12ca34 WatchSource:0}: Error finding container 0275f382fc99b37c5ce61f42a7205dce3fdc2c4efe5fefcf7d29fa8a8c12ca34: Status 404 returned error can't find the container with id 0275f382fc99b37c5ce61f42a7205dce3fdc2c4efe5fefcf7d29fa8a8c12ca34 Dec 01 09:01:01 crc kubenswrapper[4873]: I1201 09:01:01.146760 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-ld88d" event={"ID":"d94ba6ed-6043-486e-bb35-8c14add87f61","Type":"ContainerDied","Data":"5377def4e9869af146c36a76cf8d5fdd4cdea82b2c96c2d112326a1fb441f406"} Dec 01 09:01:01 crc kubenswrapper[4873]: I1201 09:01:01.146830 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5377def4e9869af146c36a76cf8d5fdd4cdea82b2c96c2d112326a1fb441f406" Dec 01 09:01:01 crc kubenswrapper[4873]: I1201 09:01:01.146945 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-ld88d" Dec 01 09:01:01 crc kubenswrapper[4873]: I1201 09:01:01.155202 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-2df0-account-create-update-k6xbx" event={"ID":"1ff3eb73-9732-4b92-a21c-3a195b39f791","Type":"ContainerDied","Data":"ce0e74560c7fb436295a3f8e6d83ee5fe2ffa8067a5c8691714a400b5a32154b"} Dec 01 09:01:01 crc kubenswrapper[4873]: I1201 09:01:01.155264 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce0e74560c7fb436295a3f8e6d83ee5fe2ffa8067a5c8691714a400b5a32154b" Dec 01 09:01:01 crc kubenswrapper[4873]: I1201 09:01:01.155333 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-2df0-account-create-update-k6xbx" Dec 01 09:01:01 crc kubenswrapper[4873]: I1201 09:01:01.164719 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"43cb2357-1bfa-40e6-8754-3b41c1e82034","Type":"ContainerStarted","Data":"2b7a01a3dce0cafa1eb41f8a0f5d371fae22727dae4c6d8741e30dde3d15d2fa"} Dec 01 09:01:01 crc kubenswrapper[4873]: I1201 09:01:01.164861 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="43cb2357-1bfa-40e6-8754-3b41c1e82034" containerName="ceilometer-central-agent" containerID="cri-o://7517972ce6cc58991973ed9d7a63f8c3d422dd492a3c0b693ce83d9ca331e879" gracePeriod=30 Dec 01 09:01:01 crc kubenswrapper[4873]: I1201 09:01:01.164996 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 09:01:01 crc kubenswrapper[4873]: I1201 09:01:01.165000 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="43cb2357-1bfa-40e6-8754-3b41c1e82034" containerName="proxy-httpd" containerID="cri-o://2b7a01a3dce0cafa1eb41f8a0f5d371fae22727dae4c6d8741e30dde3d15d2fa" gracePeriod=30 Dec 01 09:01:01 crc kubenswrapper[4873]: I1201 09:01:01.165104 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="43cb2357-1bfa-40e6-8754-3b41c1e82034" containerName="ceilometer-notification-agent" containerID="cri-o://eb3e66a154eef8f44e0931a6a8c623db26f6c343b97e67f475825e9a46c8632a" gracePeriod=30 Dec 01 09:01:01 crc kubenswrapper[4873]: I1201 09:01:01.165264 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="43cb2357-1bfa-40e6-8754-3b41c1e82034" containerName="sg-core" containerID="cri-o://972ebb4605647c3a76cc89e424ba37b80cb0950be55338668b5f488e08f760a3" gracePeriod=30 Dec 01 09:01:01 crc kubenswrapper[4873]: I1201 09:01:01.171099 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-01d7-account-create-update-rnjjk" event={"ID":"76f16e3e-d8e2-4585-aa95-61390b75713b","Type":"ContainerDied","Data":"e61896b9e53e1a39bc0ef163c5b70a472241364dbb3fad2d4672794113cc7319"} Dec 01 09:01:01 crc kubenswrapper[4873]: I1201 09:01:01.171156 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e61896b9e53e1a39bc0ef163c5b70a472241364dbb3fad2d4672794113cc7319" Dec 01 09:01:01 crc kubenswrapper[4873]: I1201 09:01:01.171237 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-01d7-account-create-update-rnjjk" Dec 01 09:01:01 crc kubenswrapper[4873]: I1201 09:01:01.197252 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.79358811 podStartE2EDuration="7.197202478s" podCreationTimestamp="2025-12-01 09:00:54 +0000 UTC" firstStartedPulling="2025-12-01 09:00:55.256540486 +0000 UTC m=+1231.158649025" lastFinishedPulling="2025-12-01 09:01:00.660154854 +0000 UTC m=+1236.562263393" observedRunningTime="2025-12-01 09:01:01.190380351 +0000 UTC m=+1237.092488880" watchObservedRunningTime="2025-12-01 09:01:01.197202478 +0000 UTC m=+1237.099311017" Dec 01 09:01:02 crc kubenswrapper[4873]: I1201 09:01:02.183491 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29409661-gls8t" event={"ID":"cf6891b3-2f06-4218-b321-5d7fac6edb7c","Type":"ContainerStarted","Data":"85b4e94c2b011ea1fe2dd65fa80d08e4f0e0969087a149e5f2adb1687d07c9d9"} Dec 01 09:01:02 crc kubenswrapper[4873]: I1201 09:01:02.184074 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29409661-gls8t" event={"ID":"cf6891b3-2f06-4218-b321-5d7fac6edb7c","Type":"ContainerStarted","Data":"0275f382fc99b37c5ce61f42a7205dce3fdc2c4efe5fefcf7d29fa8a8c12ca34"} Dec 01 09:01:02 crc kubenswrapper[4873]: I1201 09:01:02.187687 4873 generic.go:334] "Generic (PLEG): container finished" podID="43cb2357-1bfa-40e6-8754-3b41c1e82034" containerID="2b7a01a3dce0cafa1eb41f8a0f5d371fae22727dae4c6d8741e30dde3d15d2fa" exitCode=0 Dec 01 09:01:02 crc kubenswrapper[4873]: I1201 09:01:02.187722 4873 generic.go:334] "Generic (PLEG): container finished" podID="43cb2357-1bfa-40e6-8754-3b41c1e82034" containerID="972ebb4605647c3a76cc89e424ba37b80cb0950be55338668b5f488e08f760a3" exitCode=2 Dec 01 09:01:02 crc kubenswrapper[4873]: I1201 09:01:02.187730 4873 generic.go:334] "Generic (PLEG): container finished" podID="43cb2357-1bfa-40e6-8754-3b41c1e82034" containerID="eb3e66a154eef8f44e0931a6a8c623db26f6c343b97e67f475825e9a46c8632a" exitCode=0 Dec 01 09:01:02 crc kubenswrapper[4873]: I1201 09:01:02.187759 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"43cb2357-1bfa-40e6-8754-3b41c1e82034","Type":"ContainerDied","Data":"2b7a01a3dce0cafa1eb41f8a0f5d371fae22727dae4c6d8741e30dde3d15d2fa"} Dec 01 09:01:02 crc kubenswrapper[4873]: I1201 09:01:02.187788 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"43cb2357-1bfa-40e6-8754-3b41c1e82034","Type":"ContainerDied","Data":"972ebb4605647c3a76cc89e424ba37b80cb0950be55338668b5f488e08f760a3"} Dec 01 09:01:02 crc kubenswrapper[4873]: I1201 09:01:02.187801 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"43cb2357-1bfa-40e6-8754-3b41c1e82034","Type":"ContainerDied","Data":"eb3e66a154eef8f44e0931a6a8c623db26f6c343b97e67f475825e9a46c8632a"} Dec 01 09:01:02 crc kubenswrapper[4873]: I1201 09:01:02.216156 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29409661-gls8t" podStartSLOduration=2.216113426 podStartE2EDuration="2.216113426s" podCreationTimestamp="2025-12-01 09:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:01:02.202986025 +0000 UTC m=+1238.105094564" watchObservedRunningTime="2025-12-01 09:01:02.216113426 +0000 UTC m=+1238.118221975" Dec 01 09:01:04 crc kubenswrapper[4873]: I1201 09:01:04.210610 4873 generic.go:334] "Generic (PLEG): container finished" podID="cf6891b3-2f06-4218-b321-5d7fac6edb7c" containerID="85b4e94c2b011ea1fe2dd65fa80d08e4f0e0969087a149e5f2adb1687d07c9d9" exitCode=0 Dec 01 09:01:04 crc kubenswrapper[4873]: I1201 09:01:04.210778 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29409661-gls8t" event={"ID":"cf6891b3-2f06-4218-b321-5d7fac6edb7c","Type":"ContainerDied","Data":"85b4e94c2b011ea1fe2dd65fa80d08e4f0e0969087a149e5f2adb1687d07c9d9"} Dec 01 09:01:05 crc kubenswrapper[4873]: I1201 09:01:05.551453 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29409661-gls8t" Dec 01 09:01:05 crc kubenswrapper[4873]: I1201 09:01:05.631545 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf6891b3-2f06-4218-b321-5d7fac6edb7c-config-data\") pod \"cf6891b3-2f06-4218-b321-5d7fac6edb7c\" (UID: \"cf6891b3-2f06-4218-b321-5d7fac6edb7c\") " Dec 01 09:01:05 crc kubenswrapper[4873]: I1201 09:01:05.631780 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf6891b3-2f06-4218-b321-5d7fac6edb7c-combined-ca-bundle\") pod \"cf6891b3-2f06-4218-b321-5d7fac6edb7c\" (UID: \"cf6891b3-2f06-4218-b321-5d7fac6edb7c\") " Dec 01 09:01:05 crc kubenswrapper[4873]: I1201 09:01:05.631816 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q885j\" (UniqueName: \"kubernetes.io/projected/cf6891b3-2f06-4218-b321-5d7fac6edb7c-kube-api-access-q885j\") pod \"cf6891b3-2f06-4218-b321-5d7fac6edb7c\" (UID: \"cf6891b3-2f06-4218-b321-5d7fac6edb7c\") " Dec 01 09:01:05 crc kubenswrapper[4873]: I1201 09:01:05.631863 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cf6891b3-2f06-4218-b321-5d7fac6edb7c-fernet-keys\") pod \"cf6891b3-2f06-4218-b321-5d7fac6edb7c\" (UID: \"cf6891b3-2f06-4218-b321-5d7fac6edb7c\") " Dec 01 09:01:05 crc kubenswrapper[4873]: I1201 09:01:05.639858 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf6891b3-2f06-4218-b321-5d7fac6edb7c-kube-api-access-q885j" (OuterVolumeSpecName: "kube-api-access-q885j") pod "cf6891b3-2f06-4218-b321-5d7fac6edb7c" (UID: "cf6891b3-2f06-4218-b321-5d7fac6edb7c"). InnerVolumeSpecName "kube-api-access-q885j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:01:05 crc kubenswrapper[4873]: I1201 09:01:05.641929 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf6891b3-2f06-4218-b321-5d7fac6edb7c-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "cf6891b3-2f06-4218-b321-5d7fac6edb7c" (UID: "cf6891b3-2f06-4218-b321-5d7fac6edb7c"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:01:05 crc kubenswrapper[4873]: I1201 09:01:05.671457 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf6891b3-2f06-4218-b321-5d7fac6edb7c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cf6891b3-2f06-4218-b321-5d7fac6edb7c" (UID: "cf6891b3-2f06-4218-b321-5d7fac6edb7c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:01:05 crc kubenswrapper[4873]: I1201 09:01:05.690535 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf6891b3-2f06-4218-b321-5d7fac6edb7c-config-data" (OuterVolumeSpecName: "config-data") pod "cf6891b3-2f06-4218-b321-5d7fac6edb7c" (UID: "cf6891b3-2f06-4218-b321-5d7fac6edb7c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:01:05 crc kubenswrapper[4873]: I1201 09:01:05.734840 4873 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf6891b3-2f06-4218-b321-5d7fac6edb7c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:05 crc kubenswrapper[4873]: I1201 09:01:05.734910 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q885j\" (UniqueName: \"kubernetes.io/projected/cf6891b3-2f06-4218-b321-5d7fac6edb7c-kube-api-access-q885j\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:05 crc kubenswrapper[4873]: I1201 09:01:05.734928 4873 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cf6891b3-2f06-4218-b321-5d7fac6edb7c-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:05 crc kubenswrapper[4873]: I1201 09:01:05.734937 4873 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf6891b3-2f06-4218-b321-5d7fac6edb7c-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:05 crc kubenswrapper[4873]: I1201 09:01:05.965310 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.041167 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43cb2357-1bfa-40e6-8754-3b41c1e82034-config-data\") pod \"43cb2357-1bfa-40e6-8754-3b41c1e82034\" (UID: \"43cb2357-1bfa-40e6-8754-3b41c1e82034\") " Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.042141 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/43cb2357-1bfa-40e6-8754-3b41c1e82034-run-httpd\") pod \"43cb2357-1bfa-40e6-8754-3b41c1e82034\" (UID: \"43cb2357-1bfa-40e6-8754-3b41c1e82034\") " Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.042211 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43cb2357-1bfa-40e6-8754-3b41c1e82034-scripts\") pod \"43cb2357-1bfa-40e6-8754-3b41c1e82034\" (UID: \"43cb2357-1bfa-40e6-8754-3b41c1e82034\") " Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.042290 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43cb2357-1bfa-40e6-8754-3b41c1e82034-combined-ca-bundle\") pod \"43cb2357-1bfa-40e6-8754-3b41c1e82034\" (UID: \"43cb2357-1bfa-40e6-8754-3b41c1e82034\") " Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.042348 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/43cb2357-1bfa-40e6-8754-3b41c1e82034-sg-core-conf-yaml\") pod \"43cb2357-1bfa-40e6-8754-3b41c1e82034\" (UID: \"43cb2357-1bfa-40e6-8754-3b41c1e82034\") " Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.042472 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/43cb2357-1bfa-40e6-8754-3b41c1e82034-log-httpd\") pod \"43cb2357-1bfa-40e6-8754-3b41c1e82034\" (UID: \"43cb2357-1bfa-40e6-8754-3b41c1e82034\") " Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.042572 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lhhjx\" (UniqueName: \"kubernetes.io/projected/43cb2357-1bfa-40e6-8754-3b41c1e82034-kube-api-access-lhhjx\") pod \"43cb2357-1bfa-40e6-8754-3b41c1e82034\" (UID: \"43cb2357-1bfa-40e6-8754-3b41c1e82034\") " Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.042589 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43cb2357-1bfa-40e6-8754-3b41c1e82034-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "43cb2357-1bfa-40e6-8754-3b41c1e82034" (UID: "43cb2357-1bfa-40e6-8754-3b41c1e82034"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.043960 4873 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/43cb2357-1bfa-40e6-8754-3b41c1e82034-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.044130 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43cb2357-1bfa-40e6-8754-3b41c1e82034-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "43cb2357-1bfa-40e6-8754-3b41c1e82034" (UID: "43cb2357-1bfa-40e6-8754-3b41c1e82034"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.046212 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43cb2357-1bfa-40e6-8754-3b41c1e82034-scripts" (OuterVolumeSpecName: "scripts") pod "43cb2357-1bfa-40e6-8754-3b41c1e82034" (UID: "43cb2357-1bfa-40e6-8754-3b41c1e82034"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.046815 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43cb2357-1bfa-40e6-8754-3b41c1e82034-kube-api-access-lhhjx" (OuterVolumeSpecName: "kube-api-access-lhhjx") pod "43cb2357-1bfa-40e6-8754-3b41c1e82034" (UID: "43cb2357-1bfa-40e6-8754-3b41c1e82034"). InnerVolumeSpecName "kube-api-access-lhhjx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.066771 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43cb2357-1bfa-40e6-8754-3b41c1e82034-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "43cb2357-1bfa-40e6-8754-3b41c1e82034" (UID: "43cb2357-1bfa-40e6-8754-3b41c1e82034"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.108195 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43cb2357-1bfa-40e6-8754-3b41c1e82034-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "43cb2357-1bfa-40e6-8754-3b41c1e82034" (UID: "43cb2357-1bfa-40e6-8754-3b41c1e82034"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.137985 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43cb2357-1bfa-40e6-8754-3b41c1e82034-config-data" (OuterVolumeSpecName: "config-data") pod "43cb2357-1bfa-40e6-8754-3b41c1e82034" (UID: "43cb2357-1bfa-40e6-8754-3b41c1e82034"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.145642 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lhhjx\" (UniqueName: \"kubernetes.io/projected/43cb2357-1bfa-40e6-8754-3b41c1e82034-kube-api-access-lhhjx\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.145826 4873 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43cb2357-1bfa-40e6-8754-3b41c1e82034-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.145883 4873 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43cb2357-1bfa-40e6-8754-3b41c1e82034-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.145936 4873 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43cb2357-1bfa-40e6-8754-3b41c1e82034-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.145987 4873 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/43cb2357-1bfa-40e6-8754-3b41c1e82034-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.146124 4873 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/43cb2357-1bfa-40e6-8754-3b41c1e82034-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.235669 4873 generic.go:334] "Generic (PLEG): container finished" podID="43cb2357-1bfa-40e6-8754-3b41c1e82034" containerID="7517972ce6cc58991973ed9d7a63f8c3d422dd492a3c0b693ce83d9ca331e879" exitCode=0 Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.236141 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.236362 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"43cb2357-1bfa-40e6-8754-3b41c1e82034","Type":"ContainerDied","Data":"7517972ce6cc58991973ed9d7a63f8c3d422dd492a3c0b693ce83d9ca331e879"} Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.236413 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"43cb2357-1bfa-40e6-8754-3b41c1e82034","Type":"ContainerDied","Data":"fbe57d240c40eaba3f0d58754d18ee525cdca2032064f86da5ff67f123eabc69"} Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.236434 4873 scope.go:117] "RemoveContainer" containerID="2b7a01a3dce0cafa1eb41f8a0f5d371fae22727dae4c6d8741e30dde3d15d2fa" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.240999 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29409661-gls8t" event={"ID":"cf6891b3-2f06-4218-b321-5d7fac6edb7c","Type":"ContainerDied","Data":"0275f382fc99b37c5ce61f42a7205dce3fdc2c4efe5fefcf7d29fa8a8c12ca34"} Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.241064 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0275f382fc99b37c5ce61f42a7205dce3fdc2c4efe5fefcf7d29fa8a8c12ca34" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.241139 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29409661-gls8t" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.270912 4873 scope.go:117] "RemoveContainer" containerID="972ebb4605647c3a76cc89e424ba37b80cb0950be55338668b5f488e08f760a3" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.302337 4873 scope.go:117] "RemoveContainer" containerID="eb3e66a154eef8f44e0931a6a8c623db26f6c343b97e67f475825e9a46c8632a" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.309925 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.330600 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.340070 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 09:01:06 crc kubenswrapper[4873]: E1201 09:01:06.340623 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43cb2357-1bfa-40e6-8754-3b41c1e82034" containerName="sg-core" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.340645 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="43cb2357-1bfa-40e6-8754-3b41c1e82034" containerName="sg-core" Dec 01 09:01:06 crc kubenswrapper[4873]: E1201 09:01:06.340662 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d94ba6ed-6043-486e-bb35-8c14add87f61" containerName="mariadb-database-create" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.340671 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="d94ba6ed-6043-486e-bb35-8c14add87f61" containerName="mariadb-database-create" Dec 01 09:01:06 crc kubenswrapper[4873]: E1201 09:01:06.340682 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ff3eb73-9732-4b92-a21c-3a195b39f791" containerName="mariadb-account-create-update" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.340689 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ff3eb73-9732-4b92-a21c-3a195b39f791" containerName="mariadb-account-create-update" Dec 01 09:01:06 crc kubenswrapper[4873]: E1201 09:01:06.340702 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43cb2357-1bfa-40e6-8754-3b41c1e82034" containerName="ceilometer-notification-agent" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.340709 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="43cb2357-1bfa-40e6-8754-3b41c1e82034" containerName="ceilometer-notification-agent" Dec 01 09:01:06 crc kubenswrapper[4873]: E1201 09:01:06.340732 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76f16e3e-d8e2-4585-aa95-61390b75713b" containerName="mariadb-account-create-update" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.340739 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="76f16e3e-d8e2-4585-aa95-61390b75713b" containerName="mariadb-account-create-update" Dec 01 09:01:06 crc kubenswrapper[4873]: E1201 09:01:06.340755 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf6891b3-2f06-4218-b321-5d7fac6edb7c" containerName="keystone-cron" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.340765 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf6891b3-2f06-4218-b321-5d7fac6edb7c" containerName="keystone-cron" Dec 01 09:01:06 crc kubenswrapper[4873]: E1201 09:01:06.340780 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43cb2357-1bfa-40e6-8754-3b41c1e82034" containerName="ceilometer-central-agent" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.340787 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="43cb2357-1bfa-40e6-8754-3b41c1e82034" containerName="ceilometer-central-agent" Dec 01 09:01:06 crc kubenswrapper[4873]: E1201 09:01:06.340804 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43cb2357-1bfa-40e6-8754-3b41c1e82034" containerName="proxy-httpd" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.340813 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="43cb2357-1bfa-40e6-8754-3b41c1e82034" containerName="proxy-httpd" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.341044 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="43cb2357-1bfa-40e6-8754-3b41c1e82034" containerName="ceilometer-notification-agent" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.341058 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="43cb2357-1bfa-40e6-8754-3b41c1e82034" containerName="ceilometer-central-agent" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.341073 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="76f16e3e-d8e2-4585-aa95-61390b75713b" containerName="mariadb-account-create-update" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.341096 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="43cb2357-1bfa-40e6-8754-3b41c1e82034" containerName="proxy-httpd" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.341117 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf6891b3-2f06-4218-b321-5d7fac6edb7c" containerName="keystone-cron" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.341130 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="43cb2357-1bfa-40e6-8754-3b41c1e82034" containerName="sg-core" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.341138 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ff3eb73-9732-4b92-a21c-3a195b39f791" containerName="mariadb-account-create-update" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.341146 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="d94ba6ed-6043-486e-bb35-8c14add87f61" containerName="mariadb-database-create" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.348203 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.353328 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.353914 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.358758 4873 scope.go:117] "RemoveContainer" containerID="7517972ce6cc58991973ed9d7a63f8c3d422dd492a3c0b693ce83d9ca331e879" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.386564 4873 scope.go:117] "RemoveContainer" containerID="2b7a01a3dce0cafa1eb41f8a0f5d371fae22727dae4c6d8741e30dde3d15d2fa" Dec 01 09:01:06 crc kubenswrapper[4873]: E1201 09:01:06.387185 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b7a01a3dce0cafa1eb41f8a0f5d371fae22727dae4c6d8741e30dde3d15d2fa\": container with ID starting with 2b7a01a3dce0cafa1eb41f8a0f5d371fae22727dae4c6d8741e30dde3d15d2fa not found: ID does not exist" containerID="2b7a01a3dce0cafa1eb41f8a0f5d371fae22727dae4c6d8741e30dde3d15d2fa" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.387242 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b7a01a3dce0cafa1eb41f8a0f5d371fae22727dae4c6d8741e30dde3d15d2fa"} err="failed to get container status \"2b7a01a3dce0cafa1eb41f8a0f5d371fae22727dae4c6d8741e30dde3d15d2fa\": rpc error: code = NotFound desc = could not find container \"2b7a01a3dce0cafa1eb41f8a0f5d371fae22727dae4c6d8741e30dde3d15d2fa\": container with ID starting with 2b7a01a3dce0cafa1eb41f8a0f5d371fae22727dae4c6d8741e30dde3d15d2fa not found: ID does not exist" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.387276 4873 scope.go:117] "RemoveContainer" containerID="972ebb4605647c3a76cc89e424ba37b80cb0950be55338668b5f488e08f760a3" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.387454 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 09:01:06 crc kubenswrapper[4873]: E1201 09:01:06.387829 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"972ebb4605647c3a76cc89e424ba37b80cb0950be55338668b5f488e08f760a3\": container with ID starting with 972ebb4605647c3a76cc89e424ba37b80cb0950be55338668b5f488e08f760a3 not found: ID does not exist" containerID="972ebb4605647c3a76cc89e424ba37b80cb0950be55338668b5f488e08f760a3" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.387865 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"972ebb4605647c3a76cc89e424ba37b80cb0950be55338668b5f488e08f760a3"} err="failed to get container status \"972ebb4605647c3a76cc89e424ba37b80cb0950be55338668b5f488e08f760a3\": rpc error: code = NotFound desc = could not find container \"972ebb4605647c3a76cc89e424ba37b80cb0950be55338668b5f488e08f760a3\": container with ID starting with 972ebb4605647c3a76cc89e424ba37b80cb0950be55338668b5f488e08f760a3 not found: ID does not exist" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.387898 4873 scope.go:117] "RemoveContainer" containerID="eb3e66a154eef8f44e0931a6a8c623db26f6c343b97e67f475825e9a46c8632a" Dec 01 09:01:06 crc kubenswrapper[4873]: E1201 09:01:06.388213 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb3e66a154eef8f44e0931a6a8c623db26f6c343b97e67f475825e9a46c8632a\": container with ID starting with eb3e66a154eef8f44e0931a6a8c623db26f6c343b97e67f475825e9a46c8632a not found: ID does not exist" containerID="eb3e66a154eef8f44e0931a6a8c623db26f6c343b97e67f475825e9a46c8632a" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.388243 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb3e66a154eef8f44e0931a6a8c623db26f6c343b97e67f475825e9a46c8632a"} err="failed to get container status \"eb3e66a154eef8f44e0931a6a8c623db26f6c343b97e67f475825e9a46c8632a\": rpc error: code = NotFound desc = could not find container \"eb3e66a154eef8f44e0931a6a8c623db26f6c343b97e67f475825e9a46c8632a\": container with ID starting with eb3e66a154eef8f44e0931a6a8c623db26f6c343b97e67f475825e9a46c8632a not found: ID does not exist" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.388261 4873 scope.go:117] "RemoveContainer" containerID="7517972ce6cc58991973ed9d7a63f8c3d422dd492a3c0b693ce83d9ca331e879" Dec 01 09:01:06 crc kubenswrapper[4873]: E1201 09:01:06.388711 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7517972ce6cc58991973ed9d7a63f8c3d422dd492a3c0b693ce83d9ca331e879\": container with ID starting with 7517972ce6cc58991973ed9d7a63f8c3d422dd492a3c0b693ce83d9ca331e879 not found: ID does not exist" containerID="7517972ce6cc58991973ed9d7a63f8c3d422dd492a3c0b693ce83d9ca331e879" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.388861 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7517972ce6cc58991973ed9d7a63f8c3d422dd492a3c0b693ce83d9ca331e879"} err="failed to get container status \"7517972ce6cc58991973ed9d7a63f8c3d422dd492a3c0b693ce83d9ca331e879\": rpc error: code = NotFound desc = could not find container \"7517972ce6cc58991973ed9d7a63f8c3d422dd492a3c0b693ce83d9ca331e879\": container with ID starting with 7517972ce6cc58991973ed9d7a63f8c3d422dd492a3c0b693ce83d9ca331e879 not found: ID does not exist" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.443190 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43cb2357-1bfa-40e6-8754-3b41c1e82034" path="/var/lib/kubelet/pods/43cb2357-1bfa-40e6-8754-3b41c1e82034/volumes" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.454408 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wr6kp\" (UniqueName: \"kubernetes.io/projected/19ab8684-f13c-41cc-a899-dfa0f037c489-kube-api-access-wr6kp\") pod \"ceilometer-0\" (UID: \"19ab8684-f13c-41cc-a899-dfa0f037c489\") " pod="openstack/ceilometer-0" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.454467 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19ab8684-f13c-41cc-a899-dfa0f037c489-config-data\") pod \"ceilometer-0\" (UID: \"19ab8684-f13c-41cc-a899-dfa0f037c489\") " pod="openstack/ceilometer-0" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.454500 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/19ab8684-f13c-41cc-a899-dfa0f037c489-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"19ab8684-f13c-41cc-a899-dfa0f037c489\") " pod="openstack/ceilometer-0" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.454524 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/19ab8684-f13c-41cc-a899-dfa0f037c489-run-httpd\") pod \"ceilometer-0\" (UID: \"19ab8684-f13c-41cc-a899-dfa0f037c489\") " pod="openstack/ceilometer-0" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.454590 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19ab8684-f13c-41cc-a899-dfa0f037c489-scripts\") pod \"ceilometer-0\" (UID: \"19ab8684-f13c-41cc-a899-dfa0f037c489\") " pod="openstack/ceilometer-0" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.454702 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/19ab8684-f13c-41cc-a899-dfa0f037c489-log-httpd\") pod \"ceilometer-0\" (UID: \"19ab8684-f13c-41cc-a899-dfa0f037c489\") " pod="openstack/ceilometer-0" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.454758 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19ab8684-f13c-41cc-a899-dfa0f037c489-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"19ab8684-f13c-41cc-a899-dfa0f037c489\") " pod="openstack/ceilometer-0" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.557192 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/19ab8684-f13c-41cc-a899-dfa0f037c489-log-httpd\") pod \"ceilometer-0\" (UID: \"19ab8684-f13c-41cc-a899-dfa0f037c489\") " pod="openstack/ceilometer-0" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.557289 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19ab8684-f13c-41cc-a899-dfa0f037c489-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"19ab8684-f13c-41cc-a899-dfa0f037c489\") " pod="openstack/ceilometer-0" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.557683 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wr6kp\" (UniqueName: \"kubernetes.io/projected/19ab8684-f13c-41cc-a899-dfa0f037c489-kube-api-access-wr6kp\") pod \"ceilometer-0\" (UID: \"19ab8684-f13c-41cc-a899-dfa0f037c489\") " pod="openstack/ceilometer-0" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.557809 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19ab8684-f13c-41cc-a899-dfa0f037c489-config-data\") pod \"ceilometer-0\" (UID: \"19ab8684-f13c-41cc-a899-dfa0f037c489\") " pod="openstack/ceilometer-0" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.557872 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/19ab8684-f13c-41cc-a899-dfa0f037c489-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"19ab8684-f13c-41cc-a899-dfa0f037c489\") " pod="openstack/ceilometer-0" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.557903 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/19ab8684-f13c-41cc-a899-dfa0f037c489-run-httpd\") pod \"ceilometer-0\" (UID: \"19ab8684-f13c-41cc-a899-dfa0f037c489\") " pod="openstack/ceilometer-0" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.557997 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19ab8684-f13c-41cc-a899-dfa0f037c489-scripts\") pod \"ceilometer-0\" (UID: \"19ab8684-f13c-41cc-a899-dfa0f037c489\") " pod="openstack/ceilometer-0" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.558085 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/19ab8684-f13c-41cc-a899-dfa0f037c489-log-httpd\") pod \"ceilometer-0\" (UID: \"19ab8684-f13c-41cc-a899-dfa0f037c489\") " pod="openstack/ceilometer-0" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.559359 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/19ab8684-f13c-41cc-a899-dfa0f037c489-run-httpd\") pod \"ceilometer-0\" (UID: \"19ab8684-f13c-41cc-a899-dfa0f037c489\") " pod="openstack/ceilometer-0" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.563261 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19ab8684-f13c-41cc-a899-dfa0f037c489-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"19ab8684-f13c-41cc-a899-dfa0f037c489\") " pod="openstack/ceilometer-0" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.563296 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/19ab8684-f13c-41cc-a899-dfa0f037c489-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"19ab8684-f13c-41cc-a899-dfa0f037c489\") " pod="openstack/ceilometer-0" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.564373 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19ab8684-f13c-41cc-a899-dfa0f037c489-config-data\") pod \"ceilometer-0\" (UID: \"19ab8684-f13c-41cc-a899-dfa0f037c489\") " pod="openstack/ceilometer-0" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.570987 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19ab8684-f13c-41cc-a899-dfa0f037c489-scripts\") pod \"ceilometer-0\" (UID: \"19ab8684-f13c-41cc-a899-dfa0f037c489\") " pod="openstack/ceilometer-0" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.578468 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wr6kp\" (UniqueName: \"kubernetes.io/projected/19ab8684-f13c-41cc-a899-dfa0f037c489-kube-api-access-wr6kp\") pod \"ceilometer-0\" (UID: \"19ab8684-f13c-41cc-a899-dfa0f037c489\") " pod="openstack/ceilometer-0" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.677289 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.962141 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-n2m5g"] Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.964505 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-n2m5g" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.967661 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.967876 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-vlh56" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.968045 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 01 09:01:06 crc kubenswrapper[4873]: I1201 09:01:06.977358 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-n2m5g"] Dec 01 09:01:07 crc kubenswrapper[4873]: I1201 09:01:07.068638 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q65kc\" (UniqueName: \"kubernetes.io/projected/53150db8-4612-45c3-a892-8d6619a82a2f-kube-api-access-q65kc\") pod \"nova-cell0-conductor-db-sync-n2m5g\" (UID: \"53150db8-4612-45c3-a892-8d6619a82a2f\") " pod="openstack/nova-cell0-conductor-db-sync-n2m5g" Dec 01 09:01:07 crc kubenswrapper[4873]: I1201 09:01:07.068742 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53150db8-4612-45c3-a892-8d6619a82a2f-scripts\") pod \"nova-cell0-conductor-db-sync-n2m5g\" (UID: \"53150db8-4612-45c3-a892-8d6619a82a2f\") " pod="openstack/nova-cell0-conductor-db-sync-n2m5g" Dec 01 09:01:07 crc kubenswrapper[4873]: I1201 09:01:07.068837 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53150db8-4612-45c3-a892-8d6619a82a2f-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-n2m5g\" (UID: \"53150db8-4612-45c3-a892-8d6619a82a2f\") " pod="openstack/nova-cell0-conductor-db-sync-n2m5g" Dec 01 09:01:07 crc kubenswrapper[4873]: I1201 09:01:07.068915 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53150db8-4612-45c3-a892-8d6619a82a2f-config-data\") pod \"nova-cell0-conductor-db-sync-n2m5g\" (UID: \"53150db8-4612-45c3-a892-8d6619a82a2f\") " pod="openstack/nova-cell0-conductor-db-sync-n2m5g" Dec 01 09:01:07 crc kubenswrapper[4873]: I1201 09:01:07.170541 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53150db8-4612-45c3-a892-8d6619a82a2f-config-data\") pod \"nova-cell0-conductor-db-sync-n2m5g\" (UID: \"53150db8-4612-45c3-a892-8d6619a82a2f\") " pod="openstack/nova-cell0-conductor-db-sync-n2m5g" Dec 01 09:01:07 crc kubenswrapper[4873]: I1201 09:01:07.170613 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q65kc\" (UniqueName: \"kubernetes.io/projected/53150db8-4612-45c3-a892-8d6619a82a2f-kube-api-access-q65kc\") pod \"nova-cell0-conductor-db-sync-n2m5g\" (UID: \"53150db8-4612-45c3-a892-8d6619a82a2f\") " pod="openstack/nova-cell0-conductor-db-sync-n2m5g" Dec 01 09:01:07 crc kubenswrapper[4873]: I1201 09:01:07.170654 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53150db8-4612-45c3-a892-8d6619a82a2f-scripts\") pod \"nova-cell0-conductor-db-sync-n2m5g\" (UID: \"53150db8-4612-45c3-a892-8d6619a82a2f\") " pod="openstack/nova-cell0-conductor-db-sync-n2m5g" Dec 01 09:01:07 crc kubenswrapper[4873]: I1201 09:01:07.170724 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53150db8-4612-45c3-a892-8d6619a82a2f-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-n2m5g\" (UID: \"53150db8-4612-45c3-a892-8d6619a82a2f\") " pod="openstack/nova-cell0-conductor-db-sync-n2m5g" Dec 01 09:01:07 crc kubenswrapper[4873]: I1201 09:01:07.178390 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53150db8-4612-45c3-a892-8d6619a82a2f-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-n2m5g\" (UID: \"53150db8-4612-45c3-a892-8d6619a82a2f\") " pod="openstack/nova-cell0-conductor-db-sync-n2m5g" Dec 01 09:01:07 crc kubenswrapper[4873]: I1201 09:01:07.179089 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53150db8-4612-45c3-a892-8d6619a82a2f-config-data\") pod \"nova-cell0-conductor-db-sync-n2m5g\" (UID: \"53150db8-4612-45c3-a892-8d6619a82a2f\") " pod="openstack/nova-cell0-conductor-db-sync-n2m5g" Dec 01 09:01:07 crc kubenswrapper[4873]: I1201 09:01:07.198715 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 09:01:07 crc kubenswrapper[4873]: I1201 09:01:07.207603 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53150db8-4612-45c3-a892-8d6619a82a2f-scripts\") pod \"nova-cell0-conductor-db-sync-n2m5g\" (UID: \"53150db8-4612-45c3-a892-8d6619a82a2f\") " pod="openstack/nova-cell0-conductor-db-sync-n2m5g" Dec 01 09:01:07 crc kubenswrapper[4873]: I1201 09:01:07.215886 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q65kc\" (UniqueName: \"kubernetes.io/projected/53150db8-4612-45c3-a892-8d6619a82a2f-kube-api-access-q65kc\") pod \"nova-cell0-conductor-db-sync-n2m5g\" (UID: \"53150db8-4612-45c3-a892-8d6619a82a2f\") " pod="openstack/nova-cell0-conductor-db-sync-n2m5g" Dec 01 09:01:07 crc kubenswrapper[4873]: I1201 09:01:07.274766 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"19ab8684-f13c-41cc-a899-dfa0f037c489","Type":"ContainerStarted","Data":"44f8d8489e24e041d8585ee7cd71a76919dfeb75760f69c955a3a7e28590e17c"} Dec 01 09:01:07 crc kubenswrapper[4873]: I1201 09:01:07.288550 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-n2m5g" Dec 01 09:01:07 crc kubenswrapper[4873]: I1201 09:01:07.805543 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-n2m5g"] Dec 01 09:01:08 crc kubenswrapper[4873]: I1201 09:01:08.288636 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-n2m5g" event={"ID":"53150db8-4612-45c3-a892-8d6619a82a2f","Type":"ContainerStarted","Data":"824326afa87f00288b983265e92bf4e4cb859b81c6db8194347bb143f07a9541"} Dec 01 09:01:09 crc kubenswrapper[4873]: I1201 09:01:09.299827 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"19ab8684-f13c-41cc-a899-dfa0f037c489","Type":"ContainerStarted","Data":"1eaf8ed2001e54bbcafb07ecae8c7965635fb67ebee8a372d2f15f130bfdbaa1"} Dec 01 09:01:10 crc kubenswrapper[4873]: I1201 09:01:10.311592 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"19ab8684-f13c-41cc-a899-dfa0f037c489","Type":"ContainerStarted","Data":"b6ec58342546fed4de3c66f59d9599b8cae785104fc72e1330f6cc3b3c84a0dc"} Dec 01 09:01:11 crc kubenswrapper[4873]: I1201 09:01:11.323866 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"19ab8684-f13c-41cc-a899-dfa0f037c489","Type":"ContainerStarted","Data":"a424e280c3d872492739e7ba6d292ed9ecf5e4fdd1405ace2e087c5b2aafbf88"} Dec 01 09:01:15 crc kubenswrapper[4873]: I1201 09:01:15.383045 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-n2m5g" event={"ID":"53150db8-4612-45c3-a892-8d6619a82a2f","Type":"ContainerStarted","Data":"ff034402aa98a628762b457748842a51a9fcaee7130f2a41fcd880a4ea5d9545"} Dec 01 09:01:15 crc kubenswrapper[4873]: I1201 09:01:15.391912 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"19ab8684-f13c-41cc-a899-dfa0f037c489","Type":"ContainerStarted","Data":"919b9f0d2ae29a0847ecaeb8668c6ae20b274c689a47a5aab0a7fb6d8dd0afe5"} Dec 01 09:01:15 crc kubenswrapper[4873]: I1201 09:01:15.392133 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 09:01:15 crc kubenswrapper[4873]: I1201 09:01:15.421680 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-n2m5g" podStartSLOduration=2.196999421 podStartE2EDuration="9.421655194s" podCreationTimestamp="2025-12-01 09:01:06 +0000 UTC" firstStartedPulling="2025-12-01 09:01:07.825691381 +0000 UTC m=+1243.727799920" lastFinishedPulling="2025-12-01 09:01:15.050347154 +0000 UTC m=+1250.952455693" observedRunningTime="2025-12-01 09:01:15.405323695 +0000 UTC m=+1251.307432234" watchObservedRunningTime="2025-12-01 09:01:15.421655194 +0000 UTC m=+1251.323763733" Dec 01 09:01:15 crc kubenswrapper[4873]: I1201 09:01:15.432366 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.592219112 podStartE2EDuration="9.432343126s" podCreationTimestamp="2025-12-01 09:01:06 +0000 UTC" firstStartedPulling="2025-12-01 09:01:07.21022461 +0000 UTC m=+1243.112333139" lastFinishedPulling="2025-12-01 09:01:15.050348604 +0000 UTC m=+1250.952457153" observedRunningTime="2025-12-01 09:01:15.428659716 +0000 UTC m=+1251.330768275" watchObservedRunningTime="2025-12-01 09:01:15.432343126 +0000 UTC m=+1251.334451675" Dec 01 09:01:30 crc kubenswrapper[4873]: I1201 09:01:30.564774 4873 generic.go:334] "Generic (PLEG): container finished" podID="53150db8-4612-45c3-a892-8d6619a82a2f" containerID="ff034402aa98a628762b457748842a51a9fcaee7130f2a41fcd880a4ea5d9545" exitCode=0 Dec 01 09:01:30 crc kubenswrapper[4873]: I1201 09:01:30.564916 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-n2m5g" event={"ID":"53150db8-4612-45c3-a892-8d6619a82a2f","Type":"ContainerDied","Data":"ff034402aa98a628762b457748842a51a9fcaee7130f2a41fcd880a4ea5d9545"} Dec 01 09:01:31 crc kubenswrapper[4873]: I1201 09:01:31.923283 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-n2m5g" Dec 01 09:01:32 crc kubenswrapper[4873]: I1201 09:01:32.019540 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53150db8-4612-45c3-a892-8d6619a82a2f-scripts\") pod \"53150db8-4612-45c3-a892-8d6619a82a2f\" (UID: \"53150db8-4612-45c3-a892-8d6619a82a2f\") " Dec 01 09:01:32 crc kubenswrapper[4873]: I1201 09:01:32.019641 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q65kc\" (UniqueName: \"kubernetes.io/projected/53150db8-4612-45c3-a892-8d6619a82a2f-kube-api-access-q65kc\") pod \"53150db8-4612-45c3-a892-8d6619a82a2f\" (UID: \"53150db8-4612-45c3-a892-8d6619a82a2f\") " Dec 01 09:01:32 crc kubenswrapper[4873]: I1201 09:01:32.019755 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53150db8-4612-45c3-a892-8d6619a82a2f-combined-ca-bundle\") pod \"53150db8-4612-45c3-a892-8d6619a82a2f\" (UID: \"53150db8-4612-45c3-a892-8d6619a82a2f\") " Dec 01 09:01:32 crc kubenswrapper[4873]: I1201 09:01:32.019778 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53150db8-4612-45c3-a892-8d6619a82a2f-config-data\") pod \"53150db8-4612-45c3-a892-8d6619a82a2f\" (UID: \"53150db8-4612-45c3-a892-8d6619a82a2f\") " Dec 01 09:01:32 crc kubenswrapper[4873]: I1201 09:01:32.027631 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53150db8-4612-45c3-a892-8d6619a82a2f-kube-api-access-q65kc" (OuterVolumeSpecName: "kube-api-access-q65kc") pod "53150db8-4612-45c3-a892-8d6619a82a2f" (UID: "53150db8-4612-45c3-a892-8d6619a82a2f"). InnerVolumeSpecName "kube-api-access-q65kc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:01:32 crc kubenswrapper[4873]: I1201 09:01:32.033389 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53150db8-4612-45c3-a892-8d6619a82a2f-scripts" (OuterVolumeSpecName: "scripts") pod "53150db8-4612-45c3-a892-8d6619a82a2f" (UID: "53150db8-4612-45c3-a892-8d6619a82a2f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:01:32 crc kubenswrapper[4873]: I1201 09:01:32.049619 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53150db8-4612-45c3-a892-8d6619a82a2f-config-data" (OuterVolumeSpecName: "config-data") pod "53150db8-4612-45c3-a892-8d6619a82a2f" (UID: "53150db8-4612-45c3-a892-8d6619a82a2f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:01:32 crc kubenswrapper[4873]: I1201 09:01:32.052478 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53150db8-4612-45c3-a892-8d6619a82a2f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "53150db8-4612-45c3-a892-8d6619a82a2f" (UID: "53150db8-4612-45c3-a892-8d6619a82a2f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:01:32 crc kubenswrapper[4873]: I1201 09:01:32.122397 4873 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53150db8-4612-45c3-a892-8d6619a82a2f-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:32 crc kubenswrapper[4873]: I1201 09:01:32.122449 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q65kc\" (UniqueName: \"kubernetes.io/projected/53150db8-4612-45c3-a892-8d6619a82a2f-kube-api-access-q65kc\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:32 crc kubenswrapper[4873]: I1201 09:01:32.122462 4873 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53150db8-4612-45c3-a892-8d6619a82a2f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:32 crc kubenswrapper[4873]: I1201 09:01:32.122472 4873 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53150db8-4612-45c3-a892-8d6619a82a2f-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:32 crc kubenswrapper[4873]: I1201 09:01:32.592343 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-n2m5g" event={"ID":"53150db8-4612-45c3-a892-8d6619a82a2f","Type":"ContainerDied","Data":"824326afa87f00288b983265e92bf4e4cb859b81c6db8194347bb143f07a9541"} Dec 01 09:01:32 crc kubenswrapper[4873]: I1201 09:01:32.592427 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-n2m5g" Dec 01 09:01:32 crc kubenswrapper[4873]: I1201 09:01:32.592441 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="824326afa87f00288b983265e92bf4e4cb859b81c6db8194347bb143f07a9541" Dec 01 09:01:32 crc kubenswrapper[4873]: I1201 09:01:32.711992 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 01 09:01:32 crc kubenswrapper[4873]: E1201 09:01:32.712457 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53150db8-4612-45c3-a892-8d6619a82a2f" containerName="nova-cell0-conductor-db-sync" Dec 01 09:01:32 crc kubenswrapper[4873]: I1201 09:01:32.712479 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="53150db8-4612-45c3-a892-8d6619a82a2f" containerName="nova-cell0-conductor-db-sync" Dec 01 09:01:32 crc kubenswrapper[4873]: I1201 09:01:32.712682 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="53150db8-4612-45c3-a892-8d6619a82a2f" containerName="nova-cell0-conductor-db-sync" Dec 01 09:01:32 crc kubenswrapper[4873]: I1201 09:01:32.713391 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 01 09:01:32 crc kubenswrapper[4873]: I1201 09:01:32.716231 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-vlh56" Dec 01 09:01:32 crc kubenswrapper[4873]: I1201 09:01:32.717128 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 01 09:01:32 crc kubenswrapper[4873]: I1201 09:01:32.729889 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 01 09:01:32 crc kubenswrapper[4873]: I1201 09:01:32.836950 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e18f53d4-43ea-46f1-ae6f-5eeee3ac5b5d-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"e18f53d4-43ea-46f1-ae6f-5eeee3ac5b5d\") " pod="openstack/nova-cell0-conductor-0" Dec 01 09:01:32 crc kubenswrapper[4873]: I1201 09:01:32.837039 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sf9mz\" (UniqueName: \"kubernetes.io/projected/e18f53d4-43ea-46f1-ae6f-5eeee3ac5b5d-kube-api-access-sf9mz\") pod \"nova-cell0-conductor-0\" (UID: \"e18f53d4-43ea-46f1-ae6f-5eeee3ac5b5d\") " pod="openstack/nova-cell0-conductor-0" Dec 01 09:01:32 crc kubenswrapper[4873]: I1201 09:01:32.837077 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e18f53d4-43ea-46f1-ae6f-5eeee3ac5b5d-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"e18f53d4-43ea-46f1-ae6f-5eeee3ac5b5d\") " pod="openstack/nova-cell0-conductor-0" Dec 01 09:01:32 crc kubenswrapper[4873]: I1201 09:01:32.939272 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sf9mz\" (UniqueName: \"kubernetes.io/projected/e18f53d4-43ea-46f1-ae6f-5eeee3ac5b5d-kube-api-access-sf9mz\") pod \"nova-cell0-conductor-0\" (UID: \"e18f53d4-43ea-46f1-ae6f-5eeee3ac5b5d\") " pod="openstack/nova-cell0-conductor-0" Dec 01 09:01:32 crc kubenswrapper[4873]: I1201 09:01:32.939370 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e18f53d4-43ea-46f1-ae6f-5eeee3ac5b5d-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"e18f53d4-43ea-46f1-ae6f-5eeee3ac5b5d\") " pod="openstack/nova-cell0-conductor-0" Dec 01 09:01:32 crc kubenswrapper[4873]: I1201 09:01:32.939563 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e18f53d4-43ea-46f1-ae6f-5eeee3ac5b5d-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"e18f53d4-43ea-46f1-ae6f-5eeee3ac5b5d\") " pod="openstack/nova-cell0-conductor-0" Dec 01 09:01:32 crc kubenswrapper[4873]: I1201 09:01:32.944920 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e18f53d4-43ea-46f1-ae6f-5eeee3ac5b5d-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"e18f53d4-43ea-46f1-ae6f-5eeee3ac5b5d\") " pod="openstack/nova-cell0-conductor-0" Dec 01 09:01:32 crc kubenswrapper[4873]: I1201 09:01:32.945180 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e18f53d4-43ea-46f1-ae6f-5eeee3ac5b5d-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"e18f53d4-43ea-46f1-ae6f-5eeee3ac5b5d\") " pod="openstack/nova-cell0-conductor-0" Dec 01 09:01:32 crc kubenswrapper[4873]: I1201 09:01:32.964496 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sf9mz\" (UniqueName: \"kubernetes.io/projected/e18f53d4-43ea-46f1-ae6f-5eeee3ac5b5d-kube-api-access-sf9mz\") pod \"nova-cell0-conductor-0\" (UID: \"e18f53d4-43ea-46f1-ae6f-5eeee3ac5b5d\") " pod="openstack/nova-cell0-conductor-0" Dec 01 09:01:33 crc kubenswrapper[4873]: I1201 09:01:33.032068 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 01 09:01:33 crc kubenswrapper[4873]: W1201 09:01:33.486737 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode18f53d4_43ea_46f1_ae6f_5eeee3ac5b5d.slice/crio-771d416c40609d2fbc06a8b11b6507fa336a72c0386c34c91068c035e7166a5a WatchSource:0}: Error finding container 771d416c40609d2fbc06a8b11b6507fa336a72c0386c34c91068c035e7166a5a: Status 404 returned error can't find the container with id 771d416c40609d2fbc06a8b11b6507fa336a72c0386c34c91068c035e7166a5a Dec 01 09:01:33 crc kubenswrapper[4873]: I1201 09:01:33.489894 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 01 09:01:33 crc kubenswrapper[4873]: I1201 09:01:33.602150 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"e18f53d4-43ea-46f1-ae6f-5eeee3ac5b5d","Type":"ContainerStarted","Data":"771d416c40609d2fbc06a8b11b6507fa336a72c0386c34c91068c035e7166a5a"} Dec 01 09:01:34 crc kubenswrapper[4873]: I1201 09:01:34.616960 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"e18f53d4-43ea-46f1-ae6f-5eeee3ac5b5d","Type":"ContainerStarted","Data":"7ec5f4d616bf6614e90de982112619c3ccb8b51bbcb1d6f78e3c6778d080f5e8"} Dec 01 09:01:34 crc kubenswrapper[4873]: I1201 09:01:34.617704 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 01 09:01:34 crc kubenswrapper[4873]: I1201 09:01:34.646266 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.646234373 podStartE2EDuration="2.646234373s" podCreationTimestamp="2025-12-01 09:01:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:01:34.640522483 +0000 UTC m=+1270.542631032" watchObservedRunningTime="2025-12-01 09:01:34.646234373 +0000 UTC m=+1270.548342932" Dec 01 09:01:36 crc kubenswrapper[4873]: I1201 09:01:36.683138 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 01 09:01:38 crc kubenswrapper[4873]: I1201 09:01:38.062158 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 01 09:01:38 crc kubenswrapper[4873]: I1201 09:01:38.730582 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-2n4lc"] Dec 01 09:01:38 crc kubenswrapper[4873]: I1201 09:01:38.731820 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-2n4lc" Dec 01 09:01:38 crc kubenswrapper[4873]: I1201 09:01:38.735181 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 01 09:01:38 crc kubenswrapper[4873]: I1201 09:01:38.735612 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 01 09:01:38 crc kubenswrapper[4873]: I1201 09:01:38.769203 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-2n4lc"] Dec 01 09:01:38 crc kubenswrapper[4873]: I1201 09:01:38.916919 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48fbd58f-975d-45df-bcf1-7054a4b8abe3-scripts\") pod \"nova-cell0-cell-mapping-2n4lc\" (UID: \"48fbd58f-975d-45df-bcf1-7054a4b8abe3\") " pod="openstack/nova-cell0-cell-mapping-2n4lc" Dec 01 09:01:38 crc kubenswrapper[4873]: I1201 09:01:38.916996 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgl85\" (UniqueName: \"kubernetes.io/projected/48fbd58f-975d-45df-bcf1-7054a4b8abe3-kube-api-access-hgl85\") pod \"nova-cell0-cell-mapping-2n4lc\" (UID: \"48fbd58f-975d-45df-bcf1-7054a4b8abe3\") " pod="openstack/nova-cell0-cell-mapping-2n4lc" Dec 01 09:01:38 crc kubenswrapper[4873]: I1201 09:01:38.917573 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48fbd58f-975d-45df-bcf1-7054a4b8abe3-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-2n4lc\" (UID: \"48fbd58f-975d-45df-bcf1-7054a4b8abe3\") " pod="openstack/nova-cell0-cell-mapping-2n4lc" Dec 01 09:01:38 crc kubenswrapper[4873]: I1201 09:01:38.917980 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48fbd58f-975d-45df-bcf1-7054a4b8abe3-config-data\") pod \"nova-cell0-cell-mapping-2n4lc\" (UID: \"48fbd58f-975d-45df-bcf1-7054a4b8abe3\") " pod="openstack/nova-cell0-cell-mapping-2n4lc" Dec 01 09:01:38 crc kubenswrapper[4873]: I1201 09:01:38.918316 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 01 09:01:38 crc kubenswrapper[4873]: I1201 09:01:38.927827 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 09:01:38 crc kubenswrapper[4873]: I1201 09:01:38.957299 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 09:01:38 crc kubenswrapper[4873]: I1201 09:01:38.959616 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.020927 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48fbd58f-975d-45df-bcf1-7054a4b8abe3-config-data\") pod \"nova-cell0-cell-mapping-2n4lc\" (UID: \"48fbd58f-975d-45df-bcf1-7054a4b8abe3\") " pod="openstack/nova-cell0-cell-mapping-2n4lc" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.021034 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48fbd58f-975d-45df-bcf1-7054a4b8abe3-scripts\") pod \"nova-cell0-cell-mapping-2n4lc\" (UID: \"48fbd58f-975d-45df-bcf1-7054a4b8abe3\") " pod="openstack/nova-cell0-cell-mapping-2n4lc" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.021064 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgl85\" (UniqueName: \"kubernetes.io/projected/48fbd58f-975d-45df-bcf1-7054a4b8abe3-kube-api-access-hgl85\") pod \"nova-cell0-cell-mapping-2n4lc\" (UID: \"48fbd58f-975d-45df-bcf1-7054a4b8abe3\") " pod="openstack/nova-cell0-cell-mapping-2n4lc" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.021870 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48fbd58f-975d-45df-bcf1-7054a4b8abe3-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-2n4lc\" (UID: \"48fbd58f-975d-45df-bcf1-7054a4b8abe3\") " pod="openstack/nova-cell0-cell-mapping-2n4lc" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.039334 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48fbd58f-975d-45df-bcf1-7054a4b8abe3-config-data\") pod \"nova-cell0-cell-mapping-2n4lc\" (UID: \"48fbd58f-975d-45df-bcf1-7054a4b8abe3\") " pod="openstack/nova-cell0-cell-mapping-2n4lc" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.039355 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48fbd58f-975d-45df-bcf1-7054a4b8abe3-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-2n4lc\" (UID: \"48fbd58f-975d-45df-bcf1-7054a4b8abe3\") " pod="openstack/nova-cell0-cell-mapping-2n4lc" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.039570 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48fbd58f-975d-45df-bcf1-7054a4b8abe3-scripts\") pod \"nova-cell0-cell-mapping-2n4lc\" (UID: \"48fbd58f-975d-45df-bcf1-7054a4b8abe3\") " pod="openstack/nova-cell0-cell-mapping-2n4lc" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.080260 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.082066 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.082500 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgl85\" (UniqueName: \"kubernetes.io/projected/48fbd58f-975d-45df-bcf1-7054a4b8abe3-kube-api-access-hgl85\") pod \"nova-cell0-cell-mapping-2n4lc\" (UID: \"48fbd58f-975d-45df-bcf1-7054a4b8abe3\") " pod="openstack/nova-cell0-cell-mapping-2n4lc" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.095463 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.118226 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.126543 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kzsz\" (UniqueName: \"kubernetes.io/projected/a3f5400e-ce4e-405c-aa8a-9a1751f89fe1-kube-api-access-2kzsz\") pod \"nova-scheduler-0\" (UID: \"a3f5400e-ce4e-405c-aa8a-9a1751f89fe1\") " pod="openstack/nova-scheduler-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.126616 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3f5400e-ce4e-405c-aa8a-9a1751f89fe1-config-data\") pod \"nova-scheduler-0\" (UID: \"a3f5400e-ce4e-405c-aa8a-9a1751f89fe1\") " pod="openstack/nova-scheduler-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.126654 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d271d294-edf0-445a-8df2-7dd767c99d91-logs\") pod \"nova-metadata-0\" (UID: \"d271d294-edf0-445a-8df2-7dd767c99d91\") " pod="openstack/nova-metadata-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.126704 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d271d294-edf0-445a-8df2-7dd767c99d91-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d271d294-edf0-445a-8df2-7dd767c99d91\") " pod="openstack/nova-metadata-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.126932 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d271d294-edf0-445a-8df2-7dd767c99d91-config-data\") pod \"nova-metadata-0\" (UID: \"d271d294-edf0-445a-8df2-7dd767c99d91\") " pod="openstack/nova-metadata-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.126993 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3f5400e-ce4e-405c-aa8a-9a1751f89fe1-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a3f5400e-ce4e-405c-aa8a-9a1751f89fe1\") " pod="openstack/nova-scheduler-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.127220 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68rnr\" (UniqueName: \"kubernetes.io/projected/d271d294-edf0-445a-8df2-7dd767c99d91-kube-api-access-68rnr\") pod \"nova-metadata-0\" (UID: \"d271d294-edf0-445a-8df2-7dd767c99d91\") " pod="openstack/nova-metadata-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.155075 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-566b5b7845-zxctn"] Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.157442 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-566b5b7845-zxctn" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.173952 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-566b5b7845-zxctn"] Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.231918 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d271d294-edf0-445a-8df2-7dd767c99d91-config-data\") pod \"nova-metadata-0\" (UID: \"d271d294-edf0-445a-8df2-7dd767c99d91\") " pod="openstack/nova-metadata-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.231983 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3f5400e-ce4e-405c-aa8a-9a1751f89fe1-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a3f5400e-ce4e-405c-aa8a-9a1751f89fe1\") " pod="openstack/nova-scheduler-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.232057 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68rnr\" (UniqueName: \"kubernetes.io/projected/d271d294-edf0-445a-8df2-7dd767c99d91-kube-api-access-68rnr\") pod \"nova-metadata-0\" (UID: \"d271d294-edf0-445a-8df2-7dd767c99d91\") " pod="openstack/nova-metadata-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.232145 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kzsz\" (UniqueName: \"kubernetes.io/projected/a3f5400e-ce4e-405c-aa8a-9a1751f89fe1-kube-api-access-2kzsz\") pod \"nova-scheduler-0\" (UID: \"a3f5400e-ce4e-405c-aa8a-9a1751f89fe1\") " pod="openstack/nova-scheduler-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.232167 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3f5400e-ce4e-405c-aa8a-9a1751f89fe1-config-data\") pod \"nova-scheduler-0\" (UID: \"a3f5400e-ce4e-405c-aa8a-9a1751f89fe1\") " pod="openstack/nova-scheduler-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.232198 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d271d294-edf0-445a-8df2-7dd767c99d91-logs\") pod \"nova-metadata-0\" (UID: \"d271d294-edf0-445a-8df2-7dd767c99d91\") " pod="openstack/nova-metadata-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.232233 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d271d294-edf0-445a-8df2-7dd767c99d91-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d271d294-edf0-445a-8df2-7dd767c99d91\") " pod="openstack/nova-metadata-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.234159 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d271d294-edf0-445a-8df2-7dd767c99d91-logs\") pod \"nova-metadata-0\" (UID: \"d271d294-edf0-445a-8df2-7dd767c99d91\") " pod="openstack/nova-metadata-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.237922 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d271d294-edf0-445a-8df2-7dd767c99d91-config-data\") pod \"nova-metadata-0\" (UID: \"d271d294-edf0-445a-8df2-7dd767c99d91\") " pod="openstack/nova-metadata-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.241700 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d271d294-edf0-445a-8df2-7dd767c99d91-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d271d294-edf0-445a-8df2-7dd767c99d91\") " pod="openstack/nova-metadata-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.257823 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3f5400e-ce4e-405c-aa8a-9a1751f89fe1-config-data\") pod \"nova-scheduler-0\" (UID: \"a3f5400e-ce4e-405c-aa8a-9a1751f89fe1\") " pod="openstack/nova-scheduler-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.259833 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3f5400e-ce4e-405c-aa8a-9a1751f89fe1-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a3f5400e-ce4e-405c-aa8a-9a1751f89fe1\") " pod="openstack/nova-scheduler-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.259934 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.261794 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.267943 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68rnr\" (UniqueName: \"kubernetes.io/projected/d271d294-edf0-445a-8df2-7dd767c99d91-kube-api-access-68rnr\") pod \"nova-metadata-0\" (UID: \"d271d294-edf0-445a-8df2-7dd767c99d91\") " pod="openstack/nova-metadata-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.268560 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.269556 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kzsz\" (UniqueName: \"kubernetes.io/projected/a3f5400e-ce4e-405c-aa8a-9a1751f89fe1-kube-api-access-2kzsz\") pod \"nova-scheduler-0\" (UID: \"a3f5400e-ce4e-405c-aa8a-9a1751f89fe1\") " pod="openstack/nova-scheduler-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.311272 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.312899 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.330460 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.334157 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.334700 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vk596\" (UniqueName: \"kubernetes.io/projected/cd1d4f43-ecf0-4716-8491-9f322329f2c9-kube-api-access-vk596\") pod \"dnsmasq-dns-566b5b7845-zxctn\" (UID: \"cd1d4f43-ecf0-4716-8491-9f322329f2c9\") " pod="openstack/dnsmasq-dns-566b5b7845-zxctn" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.334770 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd1d4f43-ecf0-4716-8491-9f322329f2c9-config\") pod \"dnsmasq-dns-566b5b7845-zxctn\" (UID: \"cd1d4f43-ecf0-4716-8491-9f322329f2c9\") " pod="openstack/dnsmasq-dns-566b5b7845-zxctn" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.336298 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cd1d4f43-ecf0-4716-8491-9f322329f2c9-dns-svc\") pod \"dnsmasq-dns-566b5b7845-zxctn\" (UID: \"cd1d4f43-ecf0-4716-8491-9f322329f2c9\") " pod="openstack/dnsmasq-dns-566b5b7845-zxctn" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.336372 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cd1d4f43-ecf0-4716-8491-9f322329f2c9-ovsdbserver-nb\") pod \"dnsmasq-dns-566b5b7845-zxctn\" (UID: \"cd1d4f43-ecf0-4716-8491-9f322329f2c9\") " pod="openstack/dnsmasq-dns-566b5b7845-zxctn" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.336890 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cd1d4f43-ecf0-4716-8491-9f322329f2c9-ovsdbserver-sb\") pod \"dnsmasq-dns-566b5b7845-zxctn\" (UID: \"cd1d4f43-ecf0-4716-8491-9f322329f2c9\") " pod="openstack/dnsmasq-dns-566b5b7845-zxctn" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.392267 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-2n4lc" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.401096 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.439435 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8ln7\" (UniqueName: \"kubernetes.io/projected/fa610495-1013-47ed-b54d-c83962c8caf2-kube-api-access-g8ln7\") pod \"nova-cell1-novncproxy-0\" (UID: \"fa610495-1013-47ed-b54d-c83962c8caf2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.439539 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vk596\" (UniqueName: \"kubernetes.io/projected/cd1d4f43-ecf0-4716-8491-9f322329f2c9-kube-api-access-vk596\") pod \"dnsmasq-dns-566b5b7845-zxctn\" (UID: \"cd1d4f43-ecf0-4716-8491-9f322329f2c9\") " pod="openstack/dnsmasq-dns-566b5b7845-zxctn" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.439582 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd1d4f43-ecf0-4716-8491-9f322329f2c9-config\") pod \"dnsmasq-dns-566b5b7845-zxctn\" (UID: \"cd1d4f43-ecf0-4716-8491-9f322329f2c9\") " pod="openstack/dnsmasq-dns-566b5b7845-zxctn" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.439613 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cd1d4f43-ecf0-4716-8491-9f322329f2c9-dns-svc\") pod \"dnsmasq-dns-566b5b7845-zxctn\" (UID: \"cd1d4f43-ecf0-4716-8491-9f322329f2c9\") " pod="openstack/dnsmasq-dns-566b5b7845-zxctn" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.439646 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cd1d4f43-ecf0-4716-8491-9f322329f2c9-ovsdbserver-nb\") pod \"dnsmasq-dns-566b5b7845-zxctn\" (UID: \"cd1d4f43-ecf0-4716-8491-9f322329f2c9\") " pod="openstack/dnsmasq-dns-566b5b7845-zxctn" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.439702 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/488ee0a6-2ff5-4d21-807a-775375d48151-config-data\") pod \"nova-api-0\" (UID: \"488ee0a6-2ff5-4d21-807a-775375d48151\") " pod="openstack/nova-api-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.439769 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/488ee0a6-2ff5-4d21-807a-775375d48151-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"488ee0a6-2ff5-4d21-807a-775375d48151\") " pod="openstack/nova-api-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.439805 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mf5wt\" (UniqueName: \"kubernetes.io/projected/488ee0a6-2ff5-4d21-807a-775375d48151-kube-api-access-mf5wt\") pod \"nova-api-0\" (UID: \"488ee0a6-2ff5-4d21-807a-775375d48151\") " pod="openstack/nova-api-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.439839 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa610495-1013-47ed-b54d-c83962c8caf2-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"fa610495-1013-47ed-b54d-c83962c8caf2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.439996 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa610495-1013-47ed-b54d-c83962c8caf2-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"fa610495-1013-47ed-b54d-c83962c8caf2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.441401 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cd1d4f43-ecf0-4716-8491-9f322329f2c9-ovsdbserver-sb\") pod \"dnsmasq-dns-566b5b7845-zxctn\" (UID: \"cd1d4f43-ecf0-4716-8491-9f322329f2c9\") " pod="openstack/dnsmasq-dns-566b5b7845-zxctn" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.441487 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/488ee0a6-2ff5-4d21-807a-775375d48151-logs\") pod \"nova-api-0\" (UID: \"488ee0a6-2ff5-4d21-807a-775375d48151\") " pod="openstack/nova-api-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.441528 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cd1d4f43-ecf0-4716-8491-9f322329f2c9-dns-svc\") pod \"dnsmasq-dns-566b5b7845-zxctn\" (UID: \"cd1d4f43-ecf0-4716-8491-9f322329f2c9\") " pod="openstack/dnsmasq-dns-566b5b7845-zxctn" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.441884 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cd1d4f43-ecf0-4716-8491-9f322329f2c9-ovsdbserver-nb\") pod \"dnsmasq-dns-566b5b7845-zxctn\" (UID: \"cd1d4f43-ecf0-4716-8491-9f322329f2c9\") " pod="openstack/dnsmasq-dns-566b5b7845-zxctn" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.442428 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cd1d4f43-ecf0-4716-8491-9f322329f2c9-ovsdbserver-sb\") pod \"dnsmasq-dns-566b5b7845-zxctn\" (UID: \"cd1d4f43-ecf0-4716-8491-9f322329f2c9\") " pod="openstack/dnsmasq-dns-566b5b7845-zxctn" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.445645 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd1d4f43-ecf0-4716-8491-9f322329f2c9-config\") pod \"dnsmasq-dns-566b5b7845-zxctn\" (UID: \"cd1d4f43-ecf0-4716-8491-9f322329f2c9\") " pod="openstack/dnsmasq-dns-566b5b7845-zxctn" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.468067 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vk596\" (UniqueName: \"kubernetes.io/projected/cd1d4f43-ecf0-4716-8491-9f322329f2c9-kube-api-access-vk596\") pod \"dnsmasq-dns-566b5b7845-zxctn\" (UID: \"cd1d4f43-ecf0-4716-8491-9f322329f2c9\") " pod="openstack/dnsmasq-dns-566b5b7845-zxctn" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.486893 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.507089 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-566b5b7845-zxctn" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.543767 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8ln7\" (UniqueName: \"kubernetes.io/projected/fa610495-1013-47ed-b54d-c83962c8caf2-kube-api-access-g8ln7\") pod \"nova-cell1-novncproxy-0\" (UID: \"fa610495-1013-47ed-b54d-c83962c8caf2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.543907 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/488ee0a6-2ff5-4d21-807a-775375d48151-config-data\") pod \"nova-api-0\" (UID: \"488ee0a6-2ff5-4d21-807a-775375d48151\") " pod="openstack/nova-api-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.543959 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/488ee0a6-2ff5-4d21-807a-775375d48151-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"488ee0a6-2ff5-4d21-807a-775375d48151\") " pod="openstack/nova-api-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.543988 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mf5wt\" (UniqueName: \"kubernetes.io/projected/488ee0a6-2ff5-4d21-807a-775375d48151-kube-api-access-mf5wt\") pod \"nova-api-0\" (UID: \"488ee0a6-2ff5-4d21-807a-775375d48151\") " pod="openstack/nova-api-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.544049 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa610495-1013-47ed-b54d-c83962c8caf2-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"fa610495-1013-47ed-b54d-c83962c8caf2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.544081 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa610495-1013-47ed-b54d-c83962c8caf2-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"fa610495-1013-47ed-b54d-c83962c8caf2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.544117 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/488ee0a6-2ff5-4d21-807a-775375d48151-logs\") pod \"nova-api-0\" (UID: \"488ee0a6-2ff5-4d21-807a-775375d48151\") " pod="openstack/nova-api-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.548646 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/488ee0a6-2ff5-4d21-807a-775375d48151-logs\") pod \"nova-api-0\" (UID: \"488ee0a6-2ff5-4d21-807a-775375d48151\") " pod="openstack/nova-api-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.550073 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.556518 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/488ee0a6-2ff5-4d21-807a-775375d48151-config-data\") pod \"nova-api-0\" (UID: \"488ee0a6-2ff5-4d21-807a-775375d48151\") " pod="openstack/nova-api-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.557932 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa610495-1013-47ed-b54d-c83962c8caf2-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"fa610495-1013-47ed-b54d-c83962c8caf2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.558517 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/488ee0a6-2ff5-4d21-807a-775375d48151-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"488ee0a6-2ff5-4d21-807a-775375d48151\") " pod="openstack/nova-api-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.560144 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa610495-1013-47ed-b54d-c83962c8caf2-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"fa610495-1013-47ed-b54d-c83962c8caf2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.576256 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8ln7\" (UniqueName: \"kubernetes.io/projected/fa610495-1013-47ed-b54d-c83962c8caf2-kube-api-access-g8ln7\") pod \"nova-cell1-novncproxy-0\" (UID: \"fa610495-1013-47ed-b54d-c83962c8caf2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.577036 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mf5wt\" (UniqueName: \"kubernetes.io/projected/488ee0a6-2ff5-4d21-807a-775375d48151-kube-api-access-mf5wt\") pod \"nova-api-0\" (UID: \"488ee0a6-2ff5-4d21-807a-775375d48151\") " pod="openstack/nova-api-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.721254 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.732766 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.937632 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-jmbjv"] Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.940466 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-jmbjv" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.945479 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 01 09:01:39 crc kubenswrapper[4873]: I1201 09:01:39.945769 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 01 09:01:39 crc kubenswrapper[4873]: W1201 09:01:39.998338 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod48fbd58f_975d_45df_bcf1_7054a4b8abe3.slice/crio-60dbfd70398fbbe1eb3340a862ea097eb55ea06fde723e2f9437791eb1477a6e WatchSource:0}: Error finding container 60dbfd70398fbbe1eb3340a862ea097eb55ea06fde723e2f9437791eb1477a6e: Status 404 returned error can't find the container with id 60dbfd70398fbbe1eb3340a862ea097eb55ea06fde723e2f9437791eb1477a6e Dec 01 09:01:40 crc kubenswrapper[4873]: I1201 09:01:40.016323 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-jmbjv"] Dec 01 09:01:40 crc kubenswrapper[4873]: I1201 09:01:40.041626 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-2n4lc"] Dec 01 09:01:40 crc kubenswrapper[4873]: I1201 09:01:40.057661 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d-scripts\") pod \"nova-cell1-conductor-db-sync-jmbjv\" (UID: \"9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d\") " pod="openstack/nova-cell1-conductor-db-sync-jmbjv" Dec 01 09:01:40 crc kubenswrapper[4873]: I1201 09:01:40.057763 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-jmbjv\" (UID: \"9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d\") " pod="openstack/nova-cell1-conductor-db-sync-jmbjv" Dec 01 09:01:40 crc kubenswrapper[4873]: I1201 09:01:40.057831 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mscf2\" (UniqueName: \"kubernetes.io/projected/9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d-kube-api-access-mscf2\") pod \"nova-cell1-conductor-db-sync-jmbjv\" (UID: \"9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d\") " pod="openstack/nova-cell1-conductor-db-sync-jmbjv" Dec 01 09:01:40 crc kubenswrapper[4873]: I1201 09:01:40.057866 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d-config-data\") pod \"nova-cell1-conductor-db-sync-jmbjv\" (UID: \"9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d\") " pod="openstack/nova-cell1-conductor-db-sync-jmbjv" Dec 01 09:01:40 crc kubenswrapper[4873]: I1201 09:01:40.159619 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-jmbjv\" (UID: \"9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d\") " pod="openstack/nova-cell1-conductor-db-sync-jmbjv" Dec 01 09:01:40 crc kubenswrapper[4873]: I1201 09:01:40.160387 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mscf2\" (UniqueName: \"kubernetes.io/projected/9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d-kube-api-access-mscf2\") pod \"nova-cell1-conductor-db-sync-jmbjv\" (UID: \"9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d\") " pod="openstack/nova-cell1-conductor-db-sync-jmbjv" Dec 01 09:01:40 crc kubenswrapper[4873]: I1201 09:01:40.160550 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d-config-data\") pod \"nova-cell1-conductor-db-sync-jmbjv\" (UID: \"9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d\") " pod="openstack/nova-cell1-conductor-db-sync-jmbjv" Dec 01 09:01:40 crc kubenswrapper[4873]: I1201 09:01:40.160667 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d-scripts\") pod \"nova-cell1-conductor-db-sync-jmbjv\" (UID: \"9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d\") " pod="openstack/nova-cell1-conductor-db-sync-jmbjv" Dec 01 09:01:40 crc kubenswrapper[4873]: I1201 09:01:40.166823 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d-scripts\") pod \"nova-cell1-conductor-db-sync-jmbjv\" (UID: \"9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d\") " pod="openstack/nova-cell1-conductor-db-sync-jmbjv" Dec 01 09:01:40 crc kubenswrapper[4873]: I1201 09:01:40.167153 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-jmbjv\" (UID: \"9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d\") " pod="openstack/nova-cell1-conductor-db-sync-jmbjv" Dec 01 09:01:40 crc kubenswrapper[4873]: I1201 09:01:40.169191 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d-config-data\") pod \"nova-cell1-conductor-db-sync-jmbjv\" (UID: \"9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d\") " pod="openstack/nova-cell1-conductor-db-sync-jmbjv" Dec 01 09:01:40 crc kubenswrapper[4873]: I1201 09:01:40.188210 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mscf2\" (UniqueName: \"kubernetes.io/projected/9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d-kube-api-access-mscf2\") pod \"nova-cell1-conductor-db-sync-jmbjv\" (UID: \"9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d\") " pod="openstack/nova-cell1-conductor-db-sync-jmbjv" Dec 01 09:01:40 crc kubenswrapper[4873]: I1201 09:01:40.270008 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-566b5b7845-zxctn"] Dec 01 09:01:40 crc kubenswrapper[4873]: W1201 09:01:40.276813 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd1d4f43_ecf0_4716_8491_9f322329f2c9.slice/crio-d9b8c7daa1e2153464cd87447548d826c16d1677384960ba9d8673696e2bf1fd WatchSource:0}: Error finding container d9b8c7daa1e2153464cd87447548d826c16d1677384960ba9d8673696e2bf1fd: Status 404 returned error can't find the container with id d9b8c7daa1e2153464cd87447548d826c16d1677384960ba9d8673696e2bf1fd Dec 01 09:01:40 crc kubenswrapper[4873]: W1201 09:01:40.296776 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd271d294_edf0_445a_8df2_7dd767c99d91.slice/crio-002fe8ab9337f03f01c21f6c2e3004c118d8495bce9a5ea52df893c72990d090 WatchSource:0}: Error finding container 002fe8ab9337f03f01c21f6c2e3004c118d8495bce9a5ea52df893c72990d090: Status 404 returned error can't find the container with id 002fe8ab9337f03f01c21f6c2e3004c118d8495bce9a5ea52df893c72990d090 Dec 01 09:01:40 crc kubenswrapper[4873]: I1201 09:01:40.298918 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 09:01:40 crc kubenswrapper[4873]: I1201 09:01:40.304158 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-jmbjv" Dec 01 09:01:40 crc kubenswrapper[4873]: I1201 09:01:40.305037 4873 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 09:01:40 crc kubenswrapper[4873]: I1201 09:01:40.321548 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 09:01:40 crc kubenswrapper[4873]: I1201 09:01:40.460072 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 09:01:40 crc kubenswrapper[4873]: W1201 09:01:40.564989 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod488ee0a6_2ff5_4d21_807a_775375d48151.slice/crio-ad3632d91f7852372b0b8a474085c4599446547613acf55bb04129beeae5c793 WatchSource:0}: Error finding container ad3632d91f7852372b0b8a474085c4599446547613acf55bb04129beeae5c793: Status 404 returned error can't find the container with id ad3632d91f7852372b0b8a474085c4599446547613acf55bb04129beeae5c793 Dec 01 09:01:40 crc kubenswrapper[4873]: I1201 09:01:40.568063 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 09:01:40 crc kubenswrapper[4873]: I1201 09:01:40.700840 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-2n4lc" event={"ID":"48fbd58f-975d-45df-bcf1-7054a4b8abe3","Type":"ContainerStarted","Data":"4ed7f8b8d815f149f614aad11fde49b28ef841b0121fd81a73b761c17fc3e112"} Dec 01 09:01:40 crc kubenswrapper[4873]: I1201 09:01:40.700897 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-2n4lc" event={"ID":"48fbd58f-975d-45df-bcf1-7054a4b8abe3","Type":"ContainerStarted","Data":"60dbfd70398fbbe1eb3340a862ea097eb55ea06fde723e2f9437791eb1477a6e"} Dec 01 09:01:40 crc kubenswrapper[4873]: I1201 09:01:40.702888 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"fa610495-1013-47ed-b54d-c83962c8caf2","Type":"ContainerStarted","Data":"754dafd703cdbe9c70fb79bbfc368cbf1bcf907ab0b2aa0978cc1b36610a975b"} Dec 01 09:01:40 crc kubenswrapper[4873]: I1201 09:01:40.704680 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a3f5400e-ce4e-405c-aa8a-9a1751f89fe1","Type":"ContainerStarted","Data":"4e2839028511a3eea7c8a2216b67b8fcafb95b8fd45c8519125f8881ebffd312"} Dec 01 09:01:40 crc kubenswrapper[4873]: I1201 09:01:40.706781 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"488ee0a6-2ff5-4d21-807a-775375d48151","Type":"ContainerStarted","Data":"ad3632d91f7852372b0b8a474085c4599446547613acf55bb04129beeae5c793"} Dec 01 09:01:40 crc kubenswrapper[4873]: I1201 09:01:40.716227 4873 generic.go:334] "Generic (PLEG): container finished" podID="cd1d4f43-ecf0-4716-8491-9f322329f2c9" containerID="e513eeaccea770ffb3acfe6ab6c5b7450c325a0a3d6dc60e57e7ee633f39ee89" exitCode=0 Dec 01 09:01:40 crc kubenswrapper[4873]: I1201 09:01:40.716337 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-566b5b7845-zxctn" event={"ID":"cd1d4f43-ecf0-4716-8491-9f322329f2c9","Type":"ContainerDied","Data":"e513eeaccea770ffb3acfe6ab6c5b7450c325a0a3d6dc60e57e7ee633f39ee89"} Dec 01 09:01:40 crc kubenswrapper[4873]: I1201 09:01:40.716459 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-566b5b7845-zxctn" event={"ID":"cd1d4f43-ecf0-4716-8491-9f322329f2c9","Type":"ContainerStarted","Data":"d9b8c7daa1e2153464cd87447548d826c16d1677384960ba9d8673696e2bf1fd"} Dec 01 09:01:40 crc kubenswrapper[4873]: I1201 09:01:40.721490 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d271d294-edf0-445a-8df2-7dd767c99d91","Type":"ContainerStarted","Data":"002fe8ab9337f03f01c21f6c2e3004c118d8495bce9a5ea52df893c72990d090"} Dec 01 09:01:40 crc kubenswrapper[4873]: I1201 09:01:40.732831 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-2n4lc" podStartSLOduration=2.7328011930000002 podStartE2EDuration="2.732801193s" podCreationTimestamp="2025-12-01 09:01:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:01:40.716083164 +0000 UTC m=+1276.618191713" watchObservedRunningTime="2025-12-01 09:01:40.732801193 +0000 UTC m=+1276.634909742" Dec 01 09:01:40 crc kubenswrapper[4873]: I1201 09:01:40.883783 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-jmbjv"] Dec 01 09:01:41 crc kubenswrapper[4873]: I1201 09:01:41.768726 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-jmbjv" event={"ID":"9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d","Type":"ContainerStarted","Data":"5642159ae93809131ed84df354644b68602e4203c88bc6f2286dc6474f8a2232"} Dec 01 09:01:41 crc kubenswrapper[4873]: I1201 09:01:41.769259 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-jmbjv" event={"ID":"9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d","Type":"ContainerStarted","Data":"97b6f9b33b632249ab138267b59bae4b1b12d0fae188cfcad1e1cc4505929f50"} Dec 01 09:01:41 crc kubenswrapper[4873]: I1201 09:01:41.780064 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-566b5b7845-zxctn" event={"ID":"cd1d4f43-ecf0-4716-8491-9f322329f2c9","Type":"ContainerStarted","Data":"9d1768aa02375c0c2a6f07207bc68d7fbc7b61220569fa7396770fed5cb31f2c"} Dec 01 09:01:41 crc kubenswrapper[4873]: I1201 09:01:41.780121 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-566b5b7845-zxctn" Dec 01 09:01:41 crc kubenswrapper[4873]: I1201 09:01:41.793841 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-jmbjv" podStartSLOduration=2.793817621 podStartE2EDuration="2.793817621s" podCreationTimestamp="2025-12-01 09:01:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:01:41.789615418 +0000 UTC m=+1277.691723957" watchObservedRunningTime="2025-12-01 09:01:41.793817621 +0000 UTC m=+1277.695926160" Dec 01 09:01:41 crc kubenswrapper[4873]: I1201 09:01:41.824470 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-566b5b7845-zxctn" podStartSLOduration=2.82444432 podStartE2EDuration="2.82444432s" podCreationTimestamp="2025-12-01 09:01:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:01:41.820619007 +0000 UTC m=+1277.722727546" watchObservedRunningTime="2025-12-01 09:01:41.82444432 +0000 UTC m=+1277.726552859" Dec 01 09:01:42 crc kubenswrapper[4873]: I1201 09:01:42.615240 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 09:01:42 crc kubenswrapper[4873]: I1201 09:01:42.695458 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 09:01:44 crc kubenswrapper[4873]: I1201 09:01:44.019447 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 09:01:44 crc kubenswrapper[4873]: I1201 09:01:44.020252 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="c0ae1a05-ad24-4459-9a9c-1f4993efea08" containerName="kube-state-metrics" containerID="cri-o://71ffe686e209a3865545af99dca115be6ab26bd5686728be1ce81bf5da6a4107" gracePeriod=30 Dec 01 09:01:44 crc kubenswrapper[4873]: I1201 09:01:44.590129 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 09:01:44 crc kubenswrapper[4873]: I1201 09:01:44.688196 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tr5ff\" (UniqueName: \"kubernetes.io/projected/c0ae1a05-ad24-4459-9a9c-1f4993efea08-kube-api-access-tr5ff\") pod \"c0ae1a05-ad24-4459-9a9c-1f4993efea08\" (UID: \"c0ae1a05-ad24-4459-9a9c-1f4993efea08\") " Dec 01 09:01:44 crc kubenswrapper[4873]: I1201 09:01:44.695055 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0ae1a05-ad24-4459-9a9c-1f4993efea08-kube-api-access-tr5ff" (OuterVolumeSpecName: "kube-api-access-tr5ff") pod "c0ae1a05-ad24-4459-9a9c-1f4993efea08" (UID: "c0ae1a05-ad24-4459-9a9c-1f4993efea08"). InnerVolumeSpecName "kube-api-access-tr5ff". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:01:44 crc kubenswrapper[4873]: I1201 09:01:44.791430 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tr5ff\" (UniqueName: \"kubernetes.io/projected/c0ae1a05-ad24-4459-9a9c-1f4993efea08-kube-api-access-tr5ff\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:44 crc kubenswrapper[4873]: I1201 09:01:44.816108 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d271d294-edf0-445a-8df2-7dd767c99d91","Type":"ContainerStarted","Data":"5f77e4d0b56d5f8df84215b671d6f830b670a5adfa5fdedb52c448d3c18da079"} Dec 01 09:01:44 crc kubenswrapper[4873]: I1201 09:01:44.816169 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d271d294-edf0-445a-8df2-7dd767c99d91","Type":"ContainerStarted","Data":"569632beef6059c5d114c426d47b88d1797622618401a6f265c69e17ae7ed0d0"} Dec 01 09:01:44 crc kubenswrapper[4873]: I1201 09:01:44.816327 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="d271d294-edf0-445a-8df2-7dd767c99d91" containerName="nova-metadata-log" containerID="cri-o://569632beef6059c5d114c426d47b88d1797622618401a6f265c69e17ae7ed0d0" gracePeriod=30 Dec 01 09:01:44 crc kubenswrapper[4873]: I1201 09:01:44.817075 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="d271d294-edf0-445a-8df2-7dd767c99d91" containerName="nova-metadata-metadata" containerID="cri-o://5f77e4d0b56d5f8df84215b671d6f830b670a5adfa5fdedb52c448d3c18da079" gracePeriod=30 Dec 01 09:01:44 crc kubenswrapper[4873]: I1201 09:01:44.837172 4873 generic.go:334] "Generic (PLEG): container finished" podID="c0ae1a05-ad24-4459-9a9c-1f4993efea08" containerID="71ffe686e209a3865545af99dca115be6ab26bd5686728be1ce81bf5da6a4107" exitCode=2 Dec 01 09:01:44 crc kubenswrapper[4873]: I1201 09:01:44.837399 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 09:01:44 crc kubenswrapper[4873]: I1201 09:01:44.837383 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c0ae1a05-ad24-4459-9a9c-1f4993efea08","Type":"ContainerDied","Data":"71ffe686e209a3865545af99dca115be6ab26bd5686728be1ce81bf5da6a4107"} Dec 01 09:01:44 crc kubenswrapper[4873]: I1201 09:01:44.841692 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c0ae1a05-ad24-4459-9a9c-1f4993efea08","Type":"ContainerDied","Data":"65a28273c690248b4e441b31cae5ccb257f159f09dd01ecc4b8ac214beb89190"} Dec 01 09:01:44 crc kubenswrapper[4873]: I1201 09:01:44.841841 4873 scope.go:117] "RemoveContainer" containerID="71ffe686e209a3865545af99dca115be6ab26bd5686728be1ce81bf5da6a4107" Dec 01 09:01:44 crc kubenswrapper[4873]: I1201 09:01:44.846462 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"fa610495-1013-47ed-b54d-c83962c8caf2","Type":"ContainerStarted","Data":"2100353d6abdda91e019b49423ec64a465c08416e6116500b8f57695fcd32b45"} Dec 01 09:01:44 crc kubenswrapper[4873]: I1201 09:01:44.846825 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="fa610495-1013-47ed-b54d-c83962c8caf2" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://2100353d6abdda91e019b49423ec64a465c08416e6116500b8f57695fcd32b45" gracePeriod=30 Dec 01 09:01:44 crc kubenswrapper[4873]: I1201 09:01:44.869424 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.3642215650000002 podStartE2EDuration="6.869393836s" podCreationTimestamp="2025-12-01 09:01:38 +0000 UTC" firstStartedPulling="2025-12-01 09:01:40.304712144 +0000 UTC m=+1276.206820693" lastFinishedPulling="2025-12-01 09:01:43.809884425 +0000 UTC m=+1279.711992964" observedRunningTime="2025-12-01 09:01:44.869075328 +0000 UTC m=+1280.771183877" watchObservedRunningTime="2025-12-01 09:01:44.869393836 +0000 UTC m=+1280.771502375" Dec 01 09:01:44 crc kubenswrapper[4873]: I1201 09:01:44.869800 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a3f5400e-ce4e-405c-aa8a-9a1751f89fe1","Type":"ContainerStarted","Data":"fef3c34401abd0f5d59cd50b1060ddcaf4cba9139a4f8d95d8ee0118208ed803"} Dec 01 09:01:44 crc kubenswrapper[4873]: I1201 09:01:44.884084 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"488ee0a6-2ff5-4d21-807a-775375d48151","Type":"ContainerStarted","Data":"4d11339744726b9c76845472b2f90733ebfd28f6543df724571df4976f39b8ea"} Dec 01 09:01:44 crc kubenswrapper[4873]: I1201 09:01:44.885857 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"488ee0a6-2ff5-4d21-807a-775375d48151","Type":"ContainerStarted","Data":"0f587475d84c0bcc0cd7d7cac7c5d1235d8cf68292b2b0c2489bab77afa83ebf"} Dec 01 09:01:44 crc kubenswrapper[4873]: I1201 09:01:44.894331 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.600688338 podStartE2EDuration="5.894309455s" podCreationTimestamp="2025-12-01 09:01:39 +0000 UTC" firstStartedPulling="2025-12-01 09:01:40.516238327 +0000 UTC m=+1276.418346876" lastFinishedPulling="2025-12-01 09:01:43.809859454 +0000 UTC m=+1279.711967993" observedRunningTime="2025-12-01 09:01:44.892546042 +0000 UTC m=+1280.794654581" watchObservedRunningTime="2025-12-01 09:01:44.894309455 +0000 UTC m=+1280.796417994" Dec 01 09:01:44 crc kubenswrapper[4873]: I1201 09:01:44.934226 4873 scope.go:117] "RemoveContainer" containerID="71ffe686e209a3865545af99dca115be6ab26bd5686728be1ce81bf5da6a4107" Dec 01 09:01:44 crc kubenswrapper[4873]: E1201 09:01:44.938531 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71ffe686e209a3865545af99dca115be6ab26bd5686728be1ce81bf5da6a4107\": container with ID starting with 71ffe686e209a3865545af99dca115be6ab26bd5686728be1ce81bf5da6a4107 not found: ID does not exist" containerID="71ffe686e209a3865545af99dca115be6ab26bd5686728be1ce81bf5da6a4107" Dec 01 09:01:44 crc kubenswrapper[4873]: I1201 09:01:44.938608 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71ffe686e209a3865545af99dca115be6ab26bd5686728be1ce81bf5da6a4107"} err="failed to get container status \"71ffe686e209a3865545af99dca115be6ab26bd5686728be1ce81bf5da6a4107\": rpc error: code = NotFound desc = could not find container \"71ffe686e209a3865545af99dca115be6ab26bd5686728be1ce81bf5da6a4107\": container with ID starting with 71ffe686e209a3865545af99dca115be6ab26bd5686728be1ce81bf5da6a4107 not found: ID does not exist" Dec 01 09:01:44 crc kubenswrapper[4873]: I1201 09:01:44.990223 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.527059687 podStartE2EDuration="5.99019584s" podCreationTimestamp="2025-12-01 09:01:39 +0000 UTC" firstStartedPulling="2025-12-01 09:01:40.343458861 +0000 UTC m=+1276.245567390" lastFinishedPulling="2025-12-01 09:01:43.806595004 +0000 UTC m=+1279.708703543" observedRunningTime="2025-12-01 09:01:44.924409562 +0000 UTC m=+1280.826518111" watchObservedRunningTime="2025-12-01 09:01:44.99019584 +0000 UTC m=+1280.892304369" Dec 01 09:01:45 crc kubenswrapper[4873]: I1201 09:01:45.025314 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.776772515 podStartE2EDuration="6.025289849s" podCreationTimestamp="2025-12-01 09:01:39 +0000 UTC" firstStartedPulling="2025-12-01 09:01:40.572736669 +0000 UTC m=+1276.474845208" lastFinishedPulling="2025-12-01 09:01:43.821254003 +0000 UTC m=+1279.723362542" observedRunningTime="2025-12-01 09:01:45.001982759 +0000 UTC m=+1280.904091308" watchObservedRunningTime="2025-12-01 09:01:45.025289849 +0000 UTC m=+1280.927398388" Dec 01 09:01:45 crc kubenswrapper[4873]: I1201 09:01:45.050115 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 09:01:45 crc kubenswrapper[4873]: I1201 09:01:45.054567 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 09:01:45 crc kubenswrapper[4873]: I1201 09:01:45.065729 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 09:01:45 crc kubenswrapper[4873]: E1201 09:01:45.066609 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0ae1a05-ad24-4459-9a9c-1f4993efea08" containerName="kube-state-metrics" Dec 01 09:01:45 crc kubenswrapper[4873]: I1201 09:01:45.066781 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0ae1a05-ad24-4459-9a9c-1f4993efea08" containerName="kube-state-metrics" Dec 01 09:01:45 crc kubenswrapper[4873]: I1201 09:01:45.067104 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0ae1a05-ad24-4459-9a9c-1f4993efea08" containerName="kube-state-metrics" Dec 01 09:01:45 crc kubenswrapper[4873]: I1201 09:01:45.078908 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 09:01:45 crc kubenswrapper[4873]: I1201 09:01:45.088671 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 01 09:01:45 crc kubenswrapper[4873]: I1201 09:01:45.088710 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 01 09:01:45 crc kubenswrapper[4873]: I1201 09:01:45.092641 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 09:01:45 crc kubenswrapper[4873]: I1201 09:01:45.205320 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84cdb02b-04a2-4456-bd2e-48d9e9a896dd-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"84cdb02b-04a2-4456-bd2e-48d9e9a896dd\") " pod="openstack/kube-state-metrics-0" Dec 01 09:01:45 crc kubenswrapper[4873]: I1201 09:01:45.205406 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/84cdb02b-04a2-4456-bd2e-48d9e9a896dd-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"84cdb02b-04a2-4456-bd2e-48d9e9a896dd\") " pod="openstack/kube-state-metrics-0" Dec 01 09:01:45 crc kubenswrapper[4873]: I1201 09:01:45.205496 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkjp6\" (UniqueName: \"kubernetes.io/projected/84cdb02b-04a2-4456-bd2e-48d9e9a896dd-kube-api-access-jkjp6\") pod \"kube-state-metrics-0\" (UID: \"84cdb02b-04a2-4456-bd2e-48d9e9a896dd\") " pod="openstack/kube-state-metrics-0" Dec 01 09:01:45 crc kubenswrapper[4873]: I1201 09:01:45.205704 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/84cdb02b-04a2-4456-bd2e-48d9e9a896dd-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"84cdb02b-04a2-4456-bd2e-48d9e9a896dd\") " pod="openstack/kube-state-metrics-0" Dec 01 09:01:45 crc kubenswrapper[4873]: I1201 09:01:45.307991 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jkjp6\" (UniqueName: \"kubernetes.io/projected/84cdb02b-04a2-4456-bd2e-48d9e9a896dd-kube-api-access-jkjp6\") pod \"kube-state-metrics-0\" (UID: \"84cdb02b-04a2-4456-bd2e-48d9e9a896dd\") " pod="openstack/kube-state-metrics-0" Dec 01 09:01:45 crc kubenswrapper[4873]: I1201 09:01:45.308084 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/84cdb02b-04a2-4456-bd2e-48d9e9a896dd-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"84cdb02b-04a2-4456-bd2e-48d9e9a896dd\") " pod="openstack/kube-state-metrics-0" Dec 01 09:01:45 crc kubenswrapper[4873]: I1201 09:01:45.308126 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84cdb02b-04a2-4456-bd2e-48d9e9a896dd-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"84cdb02b-04a2-4456-bd2e-48d9e9a896dd\") " pod="openstack/kube-state-metrics-0" Dec 01 09:01:45 crc kubenswrapper[4873]: I1201 09:01:45.308157 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/84cdb02b-04a2-4456-bd2e-48d9e9a896dd-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"84cdb02b-04a2-4456-bd2e-48d9e9a896dd\") " pod="openstack/kube-state-metrics-0" Dec 01 09:01:45 crc kubenswrapper[4873]: I1201 09:01:45.315891 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84cdb02b-04a2-4456-bd2e-48d9e9a896dd-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"84cdb02b-04a2-4456-bd2e-48d9e9a896dd\") " pod="openstack/kube-state-metrics-0" Dec 01 09:01:45 crc kubenswrapper[4873]: I1201 09:01:45.318574 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/84cdb02b-04a2-4456-bd2e-48d9e9a896dd-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"84cdb02b-04a2-4456-bd2e-48d9e9a896dd\") " pod="openstack/kube-state-metrics-0" Dec 01 09:01:45 crc kubenswrapper[4873]: I1201 09:01:45.318683 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/84cdb02b-04a2-4456-bd2e-48d9e9a896dd-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"84cdb02b-04a2-4456-bd2e-48d9e9a896dd\") " pod="openstack/kube-state-metrics-0" Dec 01 09:01:45 crc kubenswrapper[4873]: I1201 09:01:45.332976 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkjp6\" (UniqueName: \"kubernetes.io/projected/84cdb02b-04a2-4456-bd2e-48d9e9a896dd-kube-api-access-jkjp6\") pod \"kube-state-metrics-0\" (UID: \"84cdb02b-04a2-4456-bd2e-48d9e9a896dd\") " pod="openstack/kube-state-metrics-0" Dec 01 09:01:45 crc kubenswrapper[4873]: I1201 09:01:45.415133 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 09:01:45 crc kubenswrapper[4873]: I1201 09:01:45.556773 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 09:01:45 crc kubenswrapper[4873]: I1201 09:01:45.557185 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="19ab8684-f13c-41cc-a899-dfa0f037c489" containerName="ceilometer-central-agent" containerID="cri-o://1eaf8ed2001e54bbcafb07ecae8c7965635fb67ebee8a372d2f15f130bfdbaa1" gracePeriod=30 Dec 01 09:01:45 crc kubenswrapper[4873]: I1201 09:01:45.557478 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="19ab8684-f13c-41cc-a899-dfa0f037c489" containerName="proxy-httpd" containerID="cri-o://919b9f0d2ae29a0847ecaeb8668c6ae20b274c689a47a5aab0a7fb6d8dd0afe5" gracePeriod=30 Dec 01 09:01:45 crc kubenswrapper[4873]: I1201 09:01:45.557691 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="19ab8684-f13c-41cc-a899-dfa0f037c489" containerName="sg-core" containerID="cri-o://a424e280c3d872492739e7ba6d292ed9ecf5e4fdd1405ace2e087c5b2aafbf88" gracePeriod=30 Dec 01 09:01:45 crc kubenswrapper[4873]: I1201 09:01:45.557718 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="19ab8684-f13c-41cc-a899-dfa0f037c489" containerName="ceilometer-notification-agent" containerID="cri-o://b6ec58342546fed4de3c66f59d9599b8cae785104fc72e1330f6cc3b3c84a0dc" gracePeriod=30 Dec 01 09:01:45 crc kubenswrapper[4873]: I1201 09:01:45.905629 4873 generic.go:334] "Generic (PLEG): container finished" podID="19ab8684-f13c-41cc-a899-dfa0f037c489" containerID="919b9f0d2ae29a0847ecaeb8668c6ae20b274c689a47a5aab0a7fb6d8dd0afe5" exitCode=0 Dec 01 09:01:45 crc kubenswrapper[4873]: I1201 09:01:45.908150 4873 generic.go:334] "Generic (PLEG): container finished" podID="19ab8684-f13c-41cc-a899-dfa0f037c489" containerID="a424e280c3d872492739e7ba6d292ed9ecf5e4fdd1405ace2e087c5b2aafbf88" exitCode=2 Dec 01 09:01:45 crc kubenswrapper[4873]: I1201 09:01:45.907086 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"19ab8684-f13c-41cc-a899-dfa0f037c489","Type":"ContainerDied","Data":"919b9f0d2ae29a0847ecaeb8668c6ae20b274c689a47a5aab0a7fb6d8dd0afe5"} Dec 01 09:01:45 crc kubenswrapper[4873]: I1201 09:01:45.908292 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"19ab8684-f13c-41cc-a899-dfa0f037c489","Type":"ContainerDied","Data":"a424e280c3d872492739e7ba6d292ed9ecf5e4fdd1405ace2e087c5b2aafbf88"} Dec 01 09:01:45 crc kubenswrapper[4873]: I1201 09:01:45.911487 4873 generic.go:334] "Generic (PLEG): container finished" podID="d271d294-edf0-445a-8df2-7dd767c99d91" containerID="569632beef6059c5d114c426d47b88d1797622618401a6f265c69e17ae7ed0d0" exitCode=143 Dec 01 09:01:45 crc kubenswrapper[4873]: I1201 09:01:45.911668 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d271d294-edf0-445a-8df2-7dd767c99d91","Type":"ContainerDied","Data":"569632beef6059c5d114c426d47b88d1797622618401a6f265c69e17ae7ed0d0"} Dec 01 09:01:45 crc kubenswrapper[4873]: I1201 09:01:45.916046 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 09:01:46 crc kubenswrapper[4873]: I1201 09:01:46.443729 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0ae1a05-ad24-4459-9a9c-1f4993efea08" path="/var/lib/kubelet/pods/c0ae1a05-ad24-4459-9a9c-1f4993efea08/volumes" Dec 01 09:01:46 crc kubenswrapper[4873]: I1201 09:01:46.925088 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"84cdb02b-04a2-4456-bd2e-48d9e9a896dd","Type":"ContainerStarted","Data":"5674580e04fa26111ba31ec252acf2474b866484321d9ad0f79115c511acd928"} Dec 01 09:01:46 crc kubenswrapper[4873]: I1201 09:01:46.925492 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"84cdb02b-04a2-4456-bd2e-48d9e9a896dd","Type":"ContainerStarted","Data":"7bcf9e4de159659fe97a0cd7723b9d1c6260e526547059a9d6cb7959a5a801ad"} Dec 01 09:01:46 crc kubenswrapper[4873]: I1201 09:01:46.926329 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 01 09:01:46 crc kubenswrapper[4873]: I1201 09:01:46.936603 4873 generic.go:334] "Generic (PLEG): container finished" podID="19ab8684-f13c-41cc-a899-dfa0f037c489" containerID="b6ec58342546fed4de3c66f59d9599b8cae785104fc72e1330f6cc3b3c84a0dc" exitCode=0 Dec 01 09:01:46 crc kubenswrapper[4873]: I1201 09:01:46.936638 4873 generic.go:334] "Generic (PLEG): container finished" podID="19ab8684-f13c-41cc-a899-dfa0f037c489" containerID="1eaf8ed2001e54bbcafb07ecae8c7965635fb67ebee8a372d2f15f130bfdbaa1" exitCode=0 Dec 01 09:01:46 crc kubenswrapper[4873]: I1201 09:01:46.936662 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"19ab8684-f13c-41cc-a899-dfa0f037c489","Type":"ContainerDied","Data":"b6ec58342546fed4de3c66f59d9599b8cae785104fc72e1330f6cc3b3c84a0dc"} Dec 01 09:01:46 crc kubenswrapper[4873]: I1201 09:01:46.937057 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"19ab8684-f13c-41cc-a899-dfa0f037c489","Type":"ContainerDied","Data":"1eaf8ed2001e54bbcafb07ecae8c7965635fb67ebee8a372d2f15f130bfdbaa1"} Dec 01 09:01:46 crc kubenswrapper[4873]: I1201 09:01:46.961410 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.596036521 podStartE2EDuration="1.961387036s" podCreationTimestamp="2025-12-01 09:01:45 +0000 UTC" firstStartedPulling="2025-12-01 09:01:45.919872676 +0000 UTC m=+1281.821981215" lastFinishedPulling="2025-12-01 09:01:46.285223191 +0000 UTC m=+1282.187331730" observedRunningTime="2025-12-01 09:01:46.953862212 +0000 UTC m=+1282.855970751" watchObservedRunningTime="2025-12-01 09:01:46.961387036 +0000 UTC m=+1282.863495565" Dec 01 09:01:47 crc kubenswrapper[4873]: I1201 09:01:47.273391 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 09:01:47 crc kubenswrapper[4873]: I1201 09:01:47.350854 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/19ab8684-f13c-41cc-a899-dfa0f037c489-sg-core-conf-yaml\") pod \"19ab8684-f13c-41cc-a899-dfa0f037c489\" (UID: \"19ab8684-f13c-41cc-a899-dfa0f037c489\") " Dec 01 09:01:47 crc kubenswrapper[4873]: I1201 09:01:47.350908 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/19ab8684-f13c-41cc-a899-dfa0f037c489-log-httpd\") pod \"19ab8684-f13c-41cc-a899-dfa0f037c489\" (UID: \"19ab8684-f13c-41cc-a899-dfa0f037c489\") " Dec 01 09:01:47 crc kubenswrapper[4873]: I1201 09:01:47.351145 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/19ab8684-f13c-41cc-a899-dfa0f037c489-run-httpd\") pod \"19ab8684-f13c-41cc-a899-dfa0f037c489\" (UID: \"19ab8684-f13c-41cc-a899-dfa0f037c489\") " Dec 01 09:01:47 crc kubenswrapper[4873]: I1201 09:01:47.351184 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19ab8684-f13c-41cc-a899-dfa0f037c489-combined-ca-bundle\") pod \"19ab8684-f13c-41cc-a899-dfa0f037c489\" (UID: \"19ab8684-f13c-41cc-a899-dfa0f037c489\") " Dec 01 09:01:47 crc kubenswrapper[4873]: I1201 09:01:47.351240 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wr6kp\" (UniqueName: \"kubernetes.io/projected/19ab8684-f13c-41cc-a899-dfa0f037c489-kube-api-access-wr6kp\") pod \"19ab8684-f13c-41cc-a899-dfa0f037c489\" (UID: \"19ab8684-f13c-41cc-a899-dfa0f037c489\") " Dec 01 09:01:47 crc kubenswrapper[4873]: I1201 09:01:47.351341 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19ab8684-f13c-41cc-a899-dfa0f037c489-scripts\") pod \"19ab8684-f13c-41cc-a899-dfa0f037c489\" (UID: \"19ab8684-f13c-41cc-a899-dfa0f037c489\") " Dec 01 09:01:47 crc kubenswrapper[4873]: I1201 09:01:47.351459 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19ab8684-f13c-41cc-a899-dfa0f037c489-config-data\") pod \"19ab8684-f13c-41cc-a899-dfa0f037c489\" (UID: \"19ab8684-f13c-41cc-a899-dfa0f037c489\") " Dec 01 09:01:47 crc kubenswrapper[4873]: I1201 09:01:47.352276 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19ab8684-f13c-41cc-a899-dfa0f037c489-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "19ab8684-f13c-41cc-a899-dfa0f037c489" (UID: "19ab8684-f13c-41cc-a899-dfa0f037c489"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:01:47 crc kubenswrapper[4873]: I1201 09:01:47.353563 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19ab8684-f13c-41cc-a899-dfa0f037c489-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "19ab8684-f13c-41cc-a899-dfa0f037c489" (UID: "19ab8684-f13c-41cc-a899-dfa0f037c489"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:01:47 crc kubenswrapper[4873]: I1201 09:01:47.359614 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19ab8684-f13c-41cc-a899-dfa0f037c489-scripts" (OuterVolumeSpecName: "scripts") pod "19ab8684-f13c-41cc-a899-dfa0f037c489" (UID: "19ab8684-f13c-41cc-a899-dfa0f037c489"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:01:47 crc kubenswrapper[4873]: I1201 09:01:47.370145 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19ab8684-f13c-41cc-a899-dfa0f037c489-kube-api-access-wr6kp" (OuterVolumeSpecName: "kube-api-access-wr6kp") pod "19ab8684-f13c-41cc-a899-dfa0f037c489" (UID: "19ab8684-f13c-41cc-a899-dfa0f037c489"). InnerVolumeSpecName "kube-api-access-wr6kp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:01:47 crc kubenswrapper[4873]: I1201 09:01:47.397494 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19ab8684-f13c-41cc-a899-dfa0f037c489-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "19ab8684-f13c-41cc-a899-dfa0f037c489" (UID: "19ab8684-f13c-41cc-a899-dfa0f037c489"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:01:47 crc kubenswrapper[4873]: I1201 09:01:47.451781 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19ab8684-f13c-41cc-a899-dfa0f037c489-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "19ab8684-f13c-41cc-a899-dfa0f037c489" (UID: "19ab8684-f13c-41cc-a899-dfa0f037c489"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:01:47 crc kubenswrapper[4873]: I1201 09:01:47.453881 4873 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/19ab8684-f13c-41cc-a899-dfa0f037c489-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:47 crc kubenswrapper[4873]: I1201 09:01:47.453918 4873 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19ab8684-f13c-41cc-a899-dfa0f037c489-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:47 crc kubenswrapper[4873]: I1201 09:01:47.453932 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wr6kp\" (UniqueName: \"kubernetes.io/projected/19ab8684-f13c-41cc-a899-dfa0f037c489-kube-api-access-wr6kp\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:47 crc kubenswrapper[4873]: I1201 09:01:47.453947 4873 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19ab8684-f13c-41cc-a899-dfa0f037c489-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:47 crc kubenswrapper[4873]: I1201 09:01:47.453955 4873 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/19ab8684-f13c-41cc-a899-dfa0f037c489-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:47 crc kubenswrapper[4873]: I1201 09:01:47.453967 4873 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/19ab8684-f13c-41cc-a899-dfa0f037c489-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:47 crc kubenswrapper[4873]: I1201 09:01:47.486749 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19ab8684-f13c-41cc-a899-dfa0f037c489-config-data" (OuterVolumeSpecName: "config-data") pod "19ab8684-f13c-41cc-a899-dfa0f037c489" (UID: "19ab8684-f13c-41cc-a899-dfa0f037c489"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:01:47 crc kubenswrapper[4873]: I1201 09:01:47.556399 4873 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19ab8684-f13c-41cc-a899-dfa0f037c489-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:47 crc kubenswrapper[4873]: I1201 09:01:47.950970 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 09:01:47 crc kubenswrapper[4873]: I1201 09:01:47.952457 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"19ab8684-f13c-41cc-a899-dfa0f037c489","Type":"ContainerDied","Data":"44f8d8489e24e041d8585ee7cd71a76919dfeb75760f69c955a3a7e28590e17c"} Dec 01 09:01:47 crc kubenswrapper[4873]: I1201 09:01:47.952565 4873 scope.go:117] "RemoveContainer" containerID="919b9f0d2ae29a0847ecaeb8668c6ae20b274c689a47a5aab0a7fb6d8dd0afe5" Dec 01 09:01:47 crc kubenswrapper[4873]: I1201 09:01:47.992705 4873 scope.go:117] "RemoveContainer" containerID="a424e280c3d872492739e7ba6d292ed9ecf5e4fdd1405ace2e087c5b2aafbf88" Dec 01 09:01:48 crc kubenswrapper[4873]: I1201 09:01:48.008622 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 09:01:48 crc kubenswrapper[4873]: I1201 09:01:48.022876 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 09:01:48 crc kubenswrapper[4873]: I1201 09:01:48.052212 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 09:01:48 crc kubenswrapper[4873]: E1201 09:01:48.053605 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19ab8684-f13c-41cc-a899-dfa0f037c489" containerName="sg-core" Dec 01 09:01:48 crc kubenswrapper[4873]: I1201 09:01:48.053642 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="19ab8684-f13c-41cc-a899-dfa0f037c489" containerName="sg-core" Dec 01 09:01:48 crc kubenswrapper[4873]: E1201 09:01:48.053694 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19ab8684-f13c-41cc-a899-dfa0f037c489" containerName="ceilometer-central-agent" Dec 01 09:01:48 crc kubenswrapper[4873]: I1201 09:01:48.053706 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="19ab8684-f13c-41cc-a899-dfa0f037c489" containerName="ceilometer-central-agent" Dec 01 09:01:48 crc kubenswrapper[4873]: E1201 09:01:48.053732 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19ab8684-f13c-41cc-a899-dfa0f037c489" containerName="proxy-httpd" Dec 01 09:01:48 crc kubenswrapper[4873]: I1201 09:01:48.053740 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="19ab8684-f13c-41cc-a899-dfa0f037c489" containerName="proxy-httpd" Dec 01 09:01:48 crc kubenswrapper[4873]: E1201 09:01:48.053790 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19ab8684-f13c-41cc-a899-dfa0f037c489" containerName="ceilometer-notification-agent" Dec 01 09:01:48 crc kubenswrapper[4873]: I1201 09:01:48.053798 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="19ab8684-f13c-41cc-a899-dfa0f037c489" containerName="ceilometer-notification-agent" Dec 01 09:01:48 crc kubenswrapper[4873]: I1201 09:01:48.055486 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="19ab8684-f13c-41cc-a899-dfa0f037c489" containerName="ceilometer-central-agent" Dec 01 09:01:48 crc kubenswrapper[4873]: I1201 09:01:48.055548 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="19ab8684-f13c-41cc-a899-dfa0f037c489" containerName="proxy-httpd" Dec 01 09:01:48 crc kubenswrapper[4873]: I1201 09:01:48.055560 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="19ab8684-f13c-41cc-a899-dfa0f037c489" containerName="sg-core" Dec 01 09:01:48 crc kubenswrapper[4873]: I1201 09:01:48.055579 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="19ab8684-f13c-41cc-a899-dfa0f037c489" containerName="ceilometer-notification-agent" Dec 01 09:01:48 crc kubenswrapper[4873]: I1201 09:01:48.073631 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 09:01:48 crc kubenswrapper[4873]: I1201 09:01:48.073788 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 09:01:48 crc kubenswrapper[4873]: I1201 09:01:48.077558 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 09:01:48 crc kubenswrapper[4873]: I1201 09:01:48.081993 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 01 09:01:48 crc kubenswrapper[4873]: I1201 09:01:48.082177 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 09:01:48 crc kubenswrapper[4873]: I1201 09:01:48.106314 4873 scope.go:117] "RemoveContainer" containerID="b6ec58342546fed4de3c66f59d9599b8cae785104fc72e1330f6cc3b3c84a0dc" Dec 01 09:01:48 crc kubenswrapper[4873]: I1201 09:01:48.131522 4873 scope.go:117] "RemoveContainer" containerID="1eaf8ed2001e54bbcafb07ecae8c7965635fb67ebee8a372d2f15f130bfdbaa1" Dec 01 09:01:48 crc kubenswrapper[4873]: I1201 09:01:48.170644 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-scripts\") pod \"ceilometer-0\" (UID: \"2d169f6e-9eff-40ba-bd22-d38a10f48bd5\") " pod="openstack/ceilometer-0" Dec 01 09:01:48 crc kubenswrapper[4873]: I1201 09:01:48.171053 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"2d169f6e-9eff-40ba-bd22-d38a10f48bd5\") " pod="openstack/ceilometer-0" Dec 01 09:01:48 crc kubenswrapper[4873]: I1201 09:01:48.171147 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2d169f6e-9eff-40ba-bd22-d38a10f48bd5\") " pod="openstack/ceilometer-0" Dec 01 09:01:48 crc kubenswrapper[4873]: I1201 09:01:48.171598 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-log-httpd\") pod \"ceilometer-0\" (UID: \"2d169f6e-9eff-40ba-bd22-d38a10f48bd5\") " pod="openstack/ceilometer-0" Dec 01 09:01:48 crc kubenswrapper[4873]: I1201 09:01:48.171691 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2d169f6e-9eff-40ba-bd22-d38a10f48bd5\") " pod="openstack/ceilometer-0" Dec 01 09:01:48 crc kubenswrapper[4873]: I1201 09:01:48.171723 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-config-data\") pod \"ceilometer-0\" (UID: \"2d169f6e-9eff-40ba-bd22-d38a10f48bd5\") " pod="openstack/ceilometer-0" Dec 01 09:01:48 crc kubenswrapper[4873]: I1201 09:01:48.171781 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-run-httpd\") pod \"ceilometer-0\" (UID: \"2d169f6e-9eff-40ba-bd22-d38a10f48bd5\") " pod="openstack/ceilometer-0" Dec 01 09:01:48 crc kubenswrapper[4873]: I1201 09:01:48.171824 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2hvx\" (UniqueName: \"kubernetes.io/projected/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-kube-api-access-q2hvx\") pod \"ceilometer-0\" (UID: \"2d169f6e-9eff-40ba-bd22-d38a10f48bd5\") " pod="openstack/ceilometer-0" Dec 01 09:01:48 crc kubenswrapper[4873]: I1201 09:01:48.273746 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-log-httpd\") pod \"ceilometer-0\" (UID: \"2d169f6e-9eff-40ba-bd22-d38a10f48bd5\") " pod="openstack/ceilometer-0" Dec 01 09:01:48 crc kubenswrapper[4873]: I1201 09:01:48.273814 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2d169f6e-9eff-40ba-bd22-d38a10f48bd5\") " pod="openstack/ceilometer-0" Dec 01 09:01:48 crc kubenswrapper[4873]: I1201 09:01:48.273845 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-config-data\") pod \"ceilometer-0\" (UID: \"2d169f6e-9eff-40ba-bd22-d38a10f48bd5\") " pod="openstack/ceilometer-0" Dec 01 09:01:48 crc kubenswrapper[4873]: I1201 09:01:48.273887 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-run-httpd\") pod \"ceilometer-0\" (UID: \"2d169f6e-9eff-40ba-bd22-d38a10f48bd5\") " pod="openstack/ceilometer-0" Dec 01 09:01:48 crc kubenswrapper[4873]: I1201 09:01:48.273911 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2hvx\" (UniqueName: \"kubernetes.io/projected/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-kube-api-access-q2hvx\") pod \"ceilometer-0\" (UID: \"2d169f6e-9eff-40ba-bd22-d38a10f48bd5\") " pod="openstack/ceilometer-0" Dec 01 09:01:48 crc kubenswrapper[4873]: I1201 09:01:48.273971 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-scripts\") pod \"ceilometer-0\" (UID: \"2d169f6e-9eff-40ba-bd22-d38a10f48bd5\") " pod="openstack/ceilometer-0" Dec 01 09:01:48 crc kubenswrapper[4873]: I1201 09:01:48.274031 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"2d169f6e-9eff-40ba-bd22-d38a10f48bd5\") " pod="openstack/ceilometer-0" Dec 01 09:01:48 crc kubenswrapper[4873]: I1201 09:01:48.274059 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2d169f6e-9eff-40ba-bd22-d38a10f48bd5\") " pod="openstack/ceilometer-0" Dec 01 09:01:48 crc kubenswrapper[4873]: I1201 09:01:48.275771 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-run-httpd\") pod \"ceilometer-0\" (UID: \"2d169f6e-9eff-40ba-bd22-d38a10f48bd5\") " pod="openstack/ceilometer-0" Dec 01 09:01:48 crc kubenswrapper[4873]: I1201 09:01:48.275894 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-log-httpd\") pod \"ceilometer-0\" (UID: \"2d169f6e-9eff-40ba-bd22-d38a10f48bd5\") " pod="openstack/ceilometer-0" Dec 01 09:01:48 crc kubenswrapper[4873]: I1201 09:01:48.283230 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-scripts\") pod \"ceilometer-0\" (UID: \"2d169f6e-9eff-40ba-bd22-d38a10f48bd5\") " pod="openstack/ceilometer-0" Dec 01 09:01:48 crc kubenswrapper[4873]: I1201 09:01:48.284701 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"2d169f6e-9eff-40ba-bd22-d38a10f48bd5\") " pod="openstack/ceilometer-0" Dec 01 09:01:48 crc kubenswrapper[4873]: I1201 09:01:48.288269 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-config-data\") pod \"ceilometer-0\" (UID: \"2d169f6e-9eff-40ba-bd22-d38a10f48bd5\") " pod="openstack/ceilometer-0" Dec 01 09:01:48 crc kubenswrapper[4873]: I1201 09:01:48.288719 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2d169f6e-9eff-40ba-bd22-d38a10f48bd5\") " pod="openstack/ceilometer-0" Dec 01 09:01:48 crc kubenswrapper[4873]: I1201 09:01:48.289458 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2d169f6e-9eff-40ba-bd22-d38a10f48bd5\") " pod="openstack/ceilometer-0" Dec 01 09:01:48 crc kubenswrapper[4873]: I1201 09:01:48.299812 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2hvx\" (UniqueName: \"kubernetes.io/projected/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-kube-api-access-q2hvx\") pod \"ceilometer-0\" (UID: \"2d169f6e-9eff-40ba-bd22-d38a10f48bd5\") " pod="openstack/ceilometer-0" Dec 01 09:01:48 crc kubenswrapper[4873]: I1201 09:01:48.403934 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 09:01:48 crc kubenswrapper[4873]: I1201 09:01:48.453906 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19ab8684-f13c-41cc-a899-dfa0f037c489" path="/var/lib/kubelet/pods/19ab8684-f13c-41cc-a899-dfa0f037c489/volumes" Dec 01 09:01:48 crc kubenswrapper[4873]: I1201 09:01:48.904164 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 09:01:48 crc kubenswrapper[4873]: I1201 09:01:48.963352 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d169f6e-9eff-40ba-bd22-d38a10f48bd5","Type":"ContainerStarted","Data":"8b8a768ed07ad79ec2753c8f1149d6273b521a25aa03a54180667bc61c945aec"} Dec 01 09:01:49 crc kubenswrapper[4873]: I1201 09:01:49.488180 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 01 09:01:49 crc kubenswrapper[4873]: I1201 09:01:49.488230 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 01 09:01:49 crc kubenswrapper[4873]: I1201 09:01:49.510209 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-566b5b7845-zxctn" Dec 01 09:01:49 crc kubenswrapper[4873]: I1201 09:01:49.523517 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 01 09:01:49 crc kubenswrapper[4873]: I1201 09:01:49.550692 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 01 09:01:49 crc kubenswrapper[4873]: I1201 09:01:49.550807 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 01 09:01:49 crc kubenswrapper[4873]: I1201 09:01:49.600732 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d97fcdd8f-qmg5n"] Dec 01 09:01:49 crc kubenswrapper[4873]: I1201 09:01:49.601151 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d97fcdd8f-qmg5n" podUID="a6a491d2-c38c-41ea-a1bd-ca6f6b276bde" containerName="dnsmasq-dns" containerID="cri-o://aa1da172728a3b124cb133ca83e6bf19837b0b45f09faf3ace40ef4e544a21ab" gracePeriod=10 Dec 01 09:01:49 crc kubenswrapper[4873]: I1201 09:01:49.723277 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 09:01:49 crc kubenswrapper[4873]: I1201 09:01:49.723827 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 09:01:49 crc kubenswrapper[4873]: I1201 09:01:49.734649 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 01 09:01:49 crc kubenswrapper[4873]: I1201 09:01:49.791366 4873 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d97fcdd8f-qmg5n" podUID="a6a491d2-c38c-41ea-a1bd-ca6f6b276bde" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.148:5353: connect: connection refused" Dec 01 09:01:49 crc kubenswrapper[4873]: I1201 09:01:49.973522 4873 generic.go:334] "Generic (PLEG): container finished" podID="48fbd58f-975d-45df-bcf1-7054a4b8abe3" containerID="4ed7f8b8d815f149f614aad11fde49b28ef841b0121fd81a73b761c17fc3e112" exitCode=0 Dec 01 09:01:49 crc kubenswrapper[4873]: I1201 09:01:49.973593 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-2n4lc" event={"ID":"48fbd58f-975d-45df-bcf1-7054a4b8abe3","Type":"ContainerDied","Data":"4ed7f8b8d815f149f614aad11fde49b28ef841b0121fd81a73b761c17fc3e112"} Dec 01 09:01:49 crc kubenswrapper[4873]: I1201 09:01:49.976101 4873 generic.go:334] "Generic (PLEG): container finished" podID="9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d" containerID="5642159ae93809131ed84df354644b68602e4203c88bc6f2286dc6474f8a2232" exitCode=0 Dec 01 09:01:49 crc kubenswrapper[4873]: I1201 09:01:49.976146 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-jmbjv" event={"ID":"9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d","Type":"ContainerDied","Data":"5642159ae93809131ed84df354644b68602e4203c88bc6f2286dc6474f8a2232"} Dec 01 09:01:49 crc kubenswrapper[4873]: I1201 09:01:49.978622 4873 generic.go:334] "Generic (PLEG): container finished" podID="a6a491d2-c38c-41ea-a1bd-ca6f6b276bde" containerID="aa1da172728a3b124cb133ca83e6bf19837b0b45f09faf3ace40ef4e544a21ab" exitCode=0 Dec 01 09:01:49 crc kubenswrapper[4873]: I1201 09:01:49.978914 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d97fcdd8f-qmg5n" event={"ID":"a6a491d2-c38c-41ea-a1bd-ca6f6b276bde","Type":"ContainerDied","Data":"aa1da172728a3b124cb133ca83e6bf19837b0b45f09faf3ace40ef4e544a21ab"} Dec 01 09:01:50 crc kubenswrapper[4873]: I1201 09:01:50.026700 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 01 09:01:50 crc kubenswrapper[4873]: I1201 09:01:50.148861 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d97fcdd8f-qmg5n" Dec 01 09:01:50 crc kubenswrapper[4873]: I1201 09:01:50.218193 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r67jf\" (UniqueName: \"kubernetes.io/projected/a6a491d2-c38c-41ea-a1bd-ca6f6b276bde-kube-api-access-r67jf\") pod \"a6a491d2-c38c-41ea-a1bd-ca6f6b276bde\" (UID: \"a6a491d2-c38c-41ea-a1bd-ca6f6b276bde\") " Dec 01 09:01:50 crc kubenswrapper[4873]: I1201 09:01:50.218305 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a6a491d2-c38c-41ea-a1bd-ca6f6b276bde-ovsdbserver-nb\") pod \"a6a491d2-c38c-41ea-a1bd-ca6f6b276bde\" (UID: \"a6a491d2-c38c-41ea-a1bd-ca6f6b276bde\") " Dec 01 09:01:50 crc kubenswrapper[4873]: I1201 09:01:50.218341 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a6a491d2-c38c-41ea-a1bd-ca6f6b276bde-dns-svc\") pod \"a6a491d2-c38c-41ea-a1bd-ca6f6b276bde\" (UID: \"a6a491d2-c38c-41ea-a1bd-ca6f6b276bde\") " Dec 01 09:01:50 crc kubenswrapper[4873]: I1201 09:01:50.218400 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a6a491d2-c38c-41ea-a1bd-ca6f6b276bde-ovsdbserver-sb\") pod \"a6a491d2-c38c-41ea-a1bd-ca6f6b276bde\" (UID: \"a6a491d2-c38c-41ea-a1bd-ca6f6b276bde\") " Dec 01 09:01:50 crc kubenswrapper[4873]: I1201 09:01:50.218664 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6a491d2-c38c-41ea-a1bd-ca6f6b276bde-config\") pod \"a6a491d2-c38c-41ea-a1bd-ca6f6b276bde\" (UID: \"a6a491d2-c38c-41ea-a1bd-ca6f6b276bde\") " Dec 01 09:01:50 crc kubenswrapper[4873]: I1201 09:01:50.254396 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6a491d2-c38c-41ea-a1bd-ca6f6b276bde-kube-api-access-r67jf" (OuterVolumeSpecName: "kube-api-access-r67jf") pod "a6a491d2-c38c-41ea-a1bd-ca6f6b276bde" (UID: "a6a491d2-c38c-41ea-a1bd-ca6f6b276bde"). InnerVolumeSpecName "kube-api-access-r67jf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:01:50 crc kubenswrapper[4873]: I1201 09:01:50.313909 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6a491d2-c38c-41ea-a1bd-ca6f6b276bde-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a6a491d2-c38c-41ea-a1bd-ca6f6b276bde" (UID: "a6a491d2-c38c-41ea-a1bd-ca6f6b276bde"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:01:50 crc kubenswrapper[4873]: I1201 09:01:50.321674 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r67jf\" (UniqueName: \"kubernetes.io/projected/a6a491d2-c38c-41ea-a1bd-ca6f6b276bde-kube-api-access-r67jf\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:50 crc kubenswrapper[4873]: I1201 09:01:50.321704 4873 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a6a491d2-c38c-41ea-a1bd-ca6f6b276bde-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:50 crc kubenswrapper[4873]: I1201 09:01:50.322713 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6a491d2-c38c-41ea-a1bd-ca6f6b276bde-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a6a491d2-c38c-41ea-a1bd-ca6f6b276bde" (UID: "a6a491d2-c38c-41ea-a1bd-ca6f6b276bde"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:01:50 crc kubenswrapper[4873]: I1201 09:01:50.336568 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6a491d2-c38c-41ea-a1bd-ca6f6b276bde-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a6a491d2-c38c-41ea-a1bd-ca6f6b276bde" (UID: "a6a491d2-c38c-41ea-a1bd-ca6f6b276bde"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:01:50 crc kubenswrapper[4873]: I1201 09:01:50.358715 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6a491d2-c38c-41ea-a1bd-ca6f6b276bde-config" (OuterVolumeSpecName: "config") pod "a6a491d2-c38c-41ea-a1bd-ca6f6b276bde" (UID: "a6a491d2-c38c-41ea-a1bd-ca6f6b276bde"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:01:50 crc kubenswrapper[4873]: I1201 09:01:50.425102 4873 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6a491d2-c38c-41ea-a1bd-ca6f6b276bde-config\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:50 crc kubenswrapper[4873]: I1201 09:01:50.425158 4873 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a6a491d2-c38c-41ea-a1bd-ca6f6b276bde-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:50 crc kubenswrapper[4873]: I1201 09:01:50.425178 4873 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a6a491d2-c38c-41ea-a1bd-ca6f6b276bde-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:50 crc kubenswrapper[4873]: I1201 09:01:50.806340 4873 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="488ee0a6-2ff5-4d21-807a-775375d48151" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.171:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 01 09:01:50 crc kubenswrapper[4873]: I1201 09:01:50.806435 4873 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="488ee0a6-2ff5-4d21-807a-775375d48151" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.171:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 01 09:01:51 crc kubenswrapper[4873]: I1201 09:01:51.025426 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d97fcdd8f-qmg5n" event={"ID":"a6a491d2-c38c-41ea-a1bd-ca6f6b276bde","Type":"ContainerDied","Data":"201142b7bf24cbd8e870f489422fb63ddf91d8caccf362e633e31a0ce5abe193"} Dec 01 09:01:51 crc kubenswrapper[4873]: I1201 09:01:51.025714 4873 scope.go:117] "RemoveContainer" containerID="aa1da172728a3b124cb133ca83e6bf19837b0b45f09faf3ace40ef4e544a21ab" Dec 01 09:01:51 crc kubenswrapper[4873]: I1201 09:01:51.025543 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d97fcdd8f-qmg5n" Dec 01 09:01:51 crc kubenswrapper[4873]: I1201 09:01:51.029464 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d169f6e-9eff-40ba-bd22-d38a10f48bd5","Type":"ContainerStarted","Data":"482dc025849cea6c027c087f224133403794cc1bc59ebf2b363714bff1f7756a"} Dec 01 09:01:51 crc kubenswrapper[4873]: I1201 09:01:51.137355 4873 scope.go:117] "RemoveContainer" containerID="5ee407ae1406fab0318c27c00c0b6d42c4759ae2fcd6462713d61cd472c55418" Dec 01 09:01:51 crc kubenswrapper[4873]: I1201 09:01:51.137516 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d97fcdd8f-qmg5n"] Dec 01 09:01:51 crc kubenswrapper[4873]: I1201 09:01:51.147082 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d97fcdd8f-qmg5n"] Dec 01 09:01:51 crc kubenswrapper[4873]: I1201 09:01:51.606322 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-jmbjv" Dec 01 09:01:51 crc kubenswrapper[4873]: I1201 09:01:51.648501 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-2n4lc" Dec 01 09:01:51 crc kubenswrapper[4873]: I1201 09:01:51.673754 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d-combined-ca-bundle\") pod \"9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d\" (UID: \"9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d\") " Dec 01 09:01:51 crc kubenswrapper[4873]: I1201 09:01:51.673930 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d-config-data\") pod \"9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d\" (UID: \"9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d\") " Dec 01 09:01:51 crc kubenswrapper[4873]: I1201 09:01:51.674055 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d-scripts\") pod \"9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d\" (UID: \"9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d\") " Dec 01 09:01:51 crc kubenswrapper[4873]: I1201 09:01:51.674308 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mscf2\" (UniqueName: \"kubernetes.io/projected/9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d-kube-api-access-mscf2\") pod \"9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d\" (UID: \"9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d\") " Dec 01 09:01:51 crc kubenswrapper[4873]: I1201 09:01:51.688379 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d-kube-api-access-mscf2" (OuterVolumeSpecName: "kube-api-access-mscf2") pod "9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d" (UID: "9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d"). InnerVolumeSpecName "kube-api-access-mscf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:01:51 crc kubenswrapper[4873]: I1201 09:01:51.688553 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d-scripts" (OuterVolumeSpecName: "scripts") pod "9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d" (UID: "9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:01:51 crc kubenswrapper[4873]: I1201 09:01:51.712138 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d" (UID: "9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:01:51 crc kubenswrapper[4873]: I1201 09:01:51.742316 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d-config-data" (OuterVolumeSpecName: "config-data") pod "9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d" (UID: "9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:01:51 crc kubenswrapper[4873]: I1201 09:01:51.777100 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48fbd58f-975d-45df-bcf1-7054a4b8abe3-combined-ca-bundle\") pod \"48fbd58f-975d-45df-bcf1-7054a4b8abe3\" (UID: \"48fbd58f-975d-45df-bcf1-7054a4b8abe3\") " Dec 01 09:01:51 crc kubenswrapper[4873]: I1201 09:01:51.777405 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48fbd58f-975d-45df-bcf1-7054a4b8abe3-scripts\") pod \"48fbd58f-975d-45df-bcf1-7054a4b8abe3\" (UID: \"48fbd58f-975d-45df-bcf1-7054a4b8abe3\") " Dec 01 09:01:51 crc kubenswrapper[4873]: I1201 09:01:51.777488 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hgl85\" (UniqueName: \"kubernetes.io/projected/48fbd58f-975d-45df-bcf1-7054a4b8abe3-kube-api-access-hgl85\") pod \"48fbd58f-975d-45df-bcf1-7054a4b8abe3\" (UID: \"48fbd58f-975d-45df-bcf1-7054a4b8abe3\") " Dec 01 09:01:51 crc kubenswrapper[4873]: I1201 09:01:51.777600 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48fbd58f-975d-45df-bcf1-7054a4b8abe3-config-data\") pod \"48fbd58f-975d-45df-bcf1-7054a4b8abe3\" (UID: \"48fbd58f-975d-45df-bcf1-7054a4b8abe3\") " Dec 01 09:01:51 crc kubenswrapper[4873]: I1201 09:01:51.779614 4873 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:51 crc kubenswrapper[4873]: I1201 09:01:51.779636 4873 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:51 crc kubenswrapper[4873]: I1201 09:01:51.779646 4873 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:51 crc kubenswrapper[4873]: I1201 09:01:51.779657 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mscf2\" (UniqueName: \"kubernetes.io/projected/9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d-kube-api-access-mscf2\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:51 crc kubenswrapper[4873]: I1201 09:01:51.781174 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48fbd58f-975d-45df-bcf1-7054a4b8abe3-kube-api-access-hgl85" (OuterVolumeSpecName: "kube-api-access-hgl85") pod "48fbd58f-975d-45df-bcf1-7054a4b8abe3" (UID: "48fbd58f-975d-45df-bcf1-7054a4b8abe3"). InnerVolumeSpecName "kube-api-access-hgl85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:01:51 crc kubenswrapper[4873]: I1201 09:01:51.781754 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48fbd58f-975d-45df-bcf1-7054a4b8abe3-scripts" (OuterVolumeSpecName: "scripts") pod "48fbd58f-975d-45df-bcf1-7054a4b8abe3" (UID: "48fbd58f-975d-45df-bcf1-7054a4b8abe3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:01:51 crc kubenswrapper[4873]: I1201 09:01:51.809146 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48fbd58f-975d-45df-bcf1-7054a4b8abe3-config-data" (OuterVolumeSpecName: "config-data") pod "48fbd58f-975d-45df-bcf1-7054a4b8abe3" (UID: "48fbd58f-975d-45df-bcf1-7054a4b8abe3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:01:51 crc kubenswrapper[4873]: I1201 09:01:51.810929 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48fbd58f-975d-45df-bcf1-7054a4b8abe3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "48fbd58f-975d-45df-bcf1-7054a4b8abe3" (UID: "48fbd58f-975d-45df-bcf1-7054a4b8abe3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:01:51 crc kubenswrapper[4873]: I1201 09:01:51.883229 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hgl85\" (UniqueName: \"kubernetes.io/projected/48fbd58f-975d-45df-bcf1-7054a4b8abe3-kube-api-access-hgl85\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:51 crc kubenswrapper[4873]: I1201 09:01:51.883289 4873 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48fbd58f-975d-45df-bcf1-7054a4b8abe3-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:51 crc kubenswrapper[4873]: I1201 09:01:51.883305 4873 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48fbd58f-975d-45df-bcf1-7054a4b8abe3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:51 crc kubenswrapper[4873]: I1201 09:01:51.883318 4873 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48fbd58f-975d-45df-bcf1-7054a4b8abe3-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:52 crc kubenswrapper[4873]: I1201 09:01:52.058939 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-2n4lc" event={"ID":"48fbd58f-975d-45df-bcf1-7054a4b8abe3","Type":"ContainerDied","Data":"60dbfd70398fbbe1eb3340a862ea097eb55ea06fde723e2f9437791eb1477a6e"} Dec 01 09:01:52 crc kubenswrapper[4873]: I1201 09:01:52.059295 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="60dbfd70398fbbe1eb3340a862ea097eb55ea06fde723e2f9437791eb1477a6e" Dec 01 09:01:52 crc kubenswrapper[4873]: I1201 09:01:52.059446 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-2n4lc" Dec 01 09:01:52 crc kubenswrapper[4873]: I1201 09:01:52.079673 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d169f6e-9eff-40ba-bd22-d38a10f48bd5","Type":"ContainerStarted","Data":"e77a90d91634036934dd350f3ef20bf08c149543c65b347a1517af80f56cabc4"} Dec 01 09:01:52 crc kubenswrapper[4873]: I1201 09:01:52.093696 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-jmbjv" event={"ID":"9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d","Type":"ContainerDied","Data":"97b6f9b33b632249ab138267b59bae4b1b12d0fae188cfcad1e1cc4505929f50"} Dec 01 09:01:52 crc kubenswrapper[4873]: I1201 09:01:52.093754 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="97b6f9b33b632249ab138267b59bae4b1b12d0fae188cfcad1e1cc4505929f50" Dec 01 09:01:52 crc kubenswrapper[4873]: I1201 09:01:52.093840 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-jmbjv" Dec 01 09:01:52 crc kubenswrapper[4873]: I1201 09:01:52.125369 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 01 09:01:52 crc kubenswrapper[4873]: E1201 09:01:52.125916 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48fbd58f-975d-45df-bcf1-7054a4b8abe3" containerName="nova-manage" Dec 01 09:01:52 crc kubenswrapper[4873]: I1201 09:01:52.125943 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="48fbd58f-975d-45df-bcf1-7054a4b8abe3" containerName="nova-manage" Dec 01 09:01:52 crc kubenswrapper[4873]: E1201 09:01:52.125956 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d" containerName="nova-cell1-conductor-db-sync" Dec 01 09:01:52 crc kubenswrapper[4873]: I1201 09:01:52.125964 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d" containerName="nova-cell1-conductor-db-sync" Dec 01 09:01:52 crc kubenswrapper[4873]: E1201 09:01:52.125979 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6a491d2-c38c-41ea-a1bd-ca6f6b276bde" containerName="dnsmasq-dns" Dec 01 09:01:52 crc kubenswrapper[4873]: I1201 09:01:52.125985 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6a491d2-c38c-41ea-a1bd-ca6f6b276bde" containerName="dnsmasq-dns" Dec 01 09:01:52 crc kubenswrapper[4873]: E1201 09:01:52.125997 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6a491d2-c38c-41ea-a1bd-ca6f6b276bde" containerName="init" Dec 01 09:01:52 crc kubenswrapper[4873]: I1201 09:01:52.126002 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6a491d2-c38c-41ea-a1bd-ca6f6b276bde" containerName="init" Dec 01 09:01:52 crc kubenswrapper[4873]: I1201 09:01:52.126289 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6a491d2-c38c-41ea-a1bd-ca6f6b276bde" containerName="dnsmasq-dns" Dec 01 09:01:52 crc kubenswrapper[4873]: I1201 09:01:52.126320 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d" containerName="nova-cell1-conductor-db-sync" Dec 01 09:01:52 crc kubenswrapper[4873]: I1201 09:01:52.126345 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="48fbd58f-975d-45df-bcf1-7054a4b8abe3" containerName="nova-manage" Dec 01 09:01:52 crc kubenswrapper[4873]: I1201 09:01:52.127199 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 01 09:01:52 crc kubenswrapper[4873]: I1201 09:01:52.130265 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 01 09:01:52 crc kubenswrapper[4873]: I1201 09:01:52.159360 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 01 09:01:52 crc kubenswrapper[4873]: I1201 09:01:52.190863 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgtjq\" (UniqueName: \"kubernetes.io/projected/425ae46c-8af5-46d8-bc7c-e7056aebdd34-kube-api-access-lgtjq\") pod \"nova-cell1-conductor-0\" (UID: \"425ae46c-8af5-46d8-bc7c-e7056aebdd34\") " pod="openstack/nova-cell1-conductor-0" Dec 01 09:01:52 crc kubenswrapper[4873]: I1201 09:01:52.190958 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/425ae46c-8af5-46d8-bc7c-e7056aebdd34-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"425ae46c-8af5-46d8-bc7c-e7056aebdd34\") " pod="openstack/nova-cell1-conductor-0" Dec 01 09:01:52 crc kubenswrapper[4873]: I1201 09:01:52.191038 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/425ae46c-8af5-46d8-bc7c-e7056aebdd34-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"425ae46c-8af5-46d8-bc7c-e7056aebdd34\") " pod="openstack/nova-cell1-conductor-0" Dec 01 09:01:52 crc kubenswrapper[4873]: I1201 09:01:52.300915 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/425ae46c-8af5-46d8-bc7c-e7056aebdd34-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"425ae46c-8af5-46d8-bc7c-e7056aebdd34\") " pod="openstack/nova-cell1-conductor-0" Dec 01 09:01:52 crc kubenswrapper[4873]: I1201 09:01:52.301073 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/425ae46c-8af5-46d8-bc7c-e7056aebdd34-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"425ae46c-8af5-46d8-bc7c-e7056aebdd34\") " pod="openstack/nova-cell1-conductor-0" Dec 01 09:01:52 crc kubenswrapper[4873]: I1201 09:01:52.301187 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgtjq\" (UniqueName: \"kubernetes.io/projected/425ae46c-8af5-46d8-bc7c-e7056aebdd34-kube-api-access-lgtjq\") pod \"nova-cell1-conductor-0\" (UID: \"425ae46c-8af5-46d8-bc7c-e7056aebdd34\") " pod="openstack/nova-cell1-conductor-0" Dec 01 09:01:52 crc kubenswrapper[4873]: I1201 09:01:52.315562 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/425ae46c-8af5-46d8-bc7c-e7056aebdd34-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"425ae46c-8af5-46d8-bc7c-e7056aebdd34\") " pod="openstack/nova-cell1-conductor-0" Dec 01 09:01:52 crc kubenswrapper[4873]: I1201 09:01:52.327621 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/425ae46c-8af5-46d8-bc7c-e7056aebdd34-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"425ae46c-8af5-46d8-bc7c-e7056aebdd34\") " pod="openstack/nova-cell1-conductor-0" Dec 01 09:01:52 crc kubenswrapper[4873]: I1201 09:01:52.329475 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 09:01:52 crc kubenswrapper[4873]: I1201 09:01:52.329744 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="a3f5400e-ce4e-405c-aa8a-9a1751f89fe1" containerName="nova-scheduler-scheduler" containerID="cri-o://fef3c34401abd0f5d59cd50b1060ddcaf4cba9139a4f8d95d8ee0118208ed803" gracePeriod=30 Dec 01 09:01:52 crc kubenswrapper[4873]: I1201 09:01:52.341931 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgtjq\" (UniqueName: \"kubernetes.io/projected/425ae46c-8af5-46d8-bc7c-e7056aebdd34-kube-api-access-lgtjq\") pod \"nova-cell1-conductor-0\" (UID: \"425ae46c-8af5-46d8-bc7c-e7056aebdd34\") " pod="openstack/nova-cell1-conductor-0" Dec 01 09:01:52 crc kubenswrapper[4873]: I1201 09:01:52.351635 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 09:01:52 crc kubenswrapper[4873]: I1201 09:01:52.352500 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="488ee0a6-2ff5-4d21-807a-775375d48151" containerName="nova-api-api" containerID="cri-o://4d11339744726b9c76845472b2f90733ebfd28f6543df724571df4976f39b8ea" gracePeriod=30 Dec 01 09:01:52 crc kubenswrapper[4873]: I1201 09:01:52.352843 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="488ee0a6-2ff5-4d21-807a-775375d48151" containerName="nova-api-log" containerID="cri-o://0f587475d84c0bcc0cd7d7cac7c5d1235d8cf68292b2b0c2489bab77afa83ebf" gracePeriod=30 Dec 01 09:01:52 crc kubenswrapper[4873]: I1201 09:01:52.443580 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6a491d2-c38c-41ea-a1bd-ca6f6b276bde" path="/var/lib/kubelet/pods/a6a491d2-c38c-41ea-a1bd-ca6f6b276bde/volumes" Dec 01 09:01:52 crc kubenswrapper[4873]: I1201 09:01:52.478552 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 01 09:01:53 crc kubenswrapper[4873]: I1201 09:01:53.050518 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 01 09:01:53 crc kubenswrapper[4873]: I1201 09:01:53.141677 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"425ae46c-8af5-46d8-bc7c-e7056aebdd34","Type":"ContainerStarted","Data":"d0bf83c8caa685243ae9c7725078304a7db267a66d74e9f4c371005cad72e3b9"} Dec 01 09:01:53 crc kubenswrapper[4873]: I1201 09:01:53.143053 4873 generic.go:334] "Generic (PLEG): container finished" podID="488ee0a6-2ff5-4d21-807a-775375d48151" containerID="0f587475d84c0bcc0cd7d7cac7c5d1235d8cf68292b2b0c2489bab77afa83ebf" exitCode=143 Dec 01 09:01:53 crc kubenswrapper[4873]: I1201 09:01:53.143107 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"488ee0a6-2ff5-4d21-807a-775375d48151","Type":"ContainerDied","Data":"0f587475d84c0bcc0cd7d7cac7c5d1235d8cf68292b2b0c2489bab77afa83ebf"} Dec 01 09:01:53 crc kubenswrapper[4873]: I1201 09:01:53.144562 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d169f6e-9eff-40ba-bd22-d38a10f48bd5","Type":"ContainerStarted","Data":"0bf51c0004c4ed88f9ae017b9c36b6bb3c3ffd1058b297d7eea4ff97bbe9653f"} Dec 01 09:01:54 crc kubenswrapper[4873]: I1201 09:01:54.167783 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"425ae46c-8af5-46d8-bc7c-e7056aebdd34","Type":"ContainerStarted","Data":"7c8df8971b4f3dc4cea6334d40d6c077268e1c4dd8c28f49b4308fa713b45714"} Dec 01 09:01:54 crc kubenswrapper[4873]: I1201 09:01:54.168592 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 01 09:01:54 crc kubenswrapper[4873]: I1201 09:01:54.198280 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.198257708 podStartE2EDuration="2.198257708s" podCreationTimestamp="2025-12-01 09:01:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:01:54.1872967 +0000 UTC m=+1290.089405239" watchObservedRunningTime="2025-12-01 09:01:54.198257708 +0000 UTC m=+1290.100366247" Dec 01 09:01:54 crc kubenswrapper[4873]: E1201 09:01:54.490120 4873 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="fef3c34401abd0f5d59cd50b1060ddcaf4cba9139a4f8d95d8ee0118208ed803" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 01 09:01:54 crc kubenswrapper[4873]: E1201 09:01:54.492364 4873 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="fef3c34401abd0f5d59cd50b1060ddcaf4cba9139a4f8d95d8ee0118208ed803" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 01 09:01:54 crc kubenswrapper[4873]: E1201 09:01:54.494305 4873 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="fef3c34401abd0f5d59cd50b1060ddcaf4cba9139a4f8d95d8ee0118208ed803" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 01 09:01:54 crc kubenswrapper[4873]: E1201 09:01:54.494380 4873 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="a3f5400e-ce4e-405c-aa8a-9a1751f89fe1" containerName="nova-scheduler-scheduler" Dec 01 09:01:55 crc kubenswrapper[4873]: I1201 09:01:55.184358 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d169f6e-9eff-40ba-bd22-d38a10f48bd5","Type":"ContainerStarted","Data":"809dd553e4c28663ea69e46f77c04be70e3a1947c263b42d2f0e2bae380edb10"} Dec 01 09:01:55 crc kubenswrapper[4873]: I1201 09:01:55.219300 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.149844453 podStartE2EDuration="7.219269098s" podCreationTimestamp="2025-12-01 09:01:48 +0000 UTC" firstStartedPulling="2025-12-01 09:01:48.932738718 +0000 UTC m=+1284.834847257" lastFinishedPulling="2025-12-01 09:01:54.002163363 +0000 UTC m=+1289.904271902" observedRunningTime="2025-12-01 09:01:55.20995198 +0000 UTC m=+1291.112060529" watchObservedRunningTime="2025-12-01 09:01:55.219269098 +0000 UTC m=+1291.121377637" Dec 01 09:01:55 crc kubenswrapper[4873]: I1201 09:01:55.425610 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 01 09:01:56 crc kubenswrapper[4873]: I1201 09:01:56.204571 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.114423 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.215645 4873 generic.go:334] "Generic (PLEG): container finished" podID="a3f5400e-ce4e-405c-aa8a-9a1751f89fe1" containerID="fef3c34401abd0f5d59cd50b1060ddcaf4cba9139a4f8d95d8ee0118208ed803" exitCode=0 Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.215794 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.215818 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a3f5400e-ce4e-405c-aa8a-9a1751f89fe1","Type":"ContainerDied","Data":"fef3c34401abd0f5d59cd50b1060ddcaf4cba9139a4f8d95d8ee0118208ed803"} Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.215894 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a3f5400e-ce4e-405c-aa8a-9a1751f89fe1","Type":"ContainerDied","Data":"4e2839028511a3eea7c8a2216b67b8fcafb95b8fd45c8519125f8881ebffd312"} Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.215927 4873 scope.go:117] "RemoveContainer" containerID="fef3c34401abd0f5d59cd50b1060ddcaf4cba9139a4f8d95d8ee0118208ed803" Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.233088 4873 generic.go:334] "Generic (PLEG): container finished" podID="488ee0a6-2ff5-4d21-807a-775375d48151" containerID="4d11339744726b9c76845472b2f90733ebfd28f6543df724571df4976f39b8ea" exitCode=0 Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.234165 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"488ee0a6-2ff5-4d21-807a-775375d48151","Type":"ContainerDied","Data":"4d11339744726b9c76845472b2f90733ebfd28f6543df724571df4976f39b8ea"} Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.235112 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3f5400e-ce4e-405c-aa8a-9a1751f89fe1-config-data\") pod \"a3f5400e-ce4e-405c-aa8a-9a1751f89fe1\" (UID: \"a3f5400e-ce4e-405c-aa8a-9a1751f89fe1\") " Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.235236 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3f5400e-ce4e-405c-aa8a-9a1751f89fe1-combined-ca-bundle\") pod \"a3f5400e-ce4e-405c-aa8a-9a1751f89fe1\" (UID: \"a3f5400e-ce4e-405c-aa8a-9a1751f89fe1\") " Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.235399 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2kzsz\" (UniqueName: \"kubernetes.io/projected/a3f5400e-ce4e-405c-aa8a-9a1751f89fe1-kube-api-access-2kzsz\") pod \"a3f5400e-ce4e-405c-aa8a-9a1751f89fe1\" (UID: \"a3f5400e-ce4e-405c-aa8a-9a1751f89fe1\") " Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.243514 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3f5400e-ce4e-405c-aa8a-9a1751f89fe1-kube-api-access-2kzsz" (OuterVolumeSpecName: "kube-api-access-2kzsz") pod "a3f5400e-ce4e-405c-aa8a-9a1751f89fe1" (UID: "a3f5400e-ce4e-405c-aa8a-9a1751f89fe1"). InnerVolumeSpecName "kube-api-access-2kzsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.267770 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3f5400e-ce4e-405c-aa8a-9a1751f89fe1-config-data" (OuterVolumeSpecName: "config-data") pod "a3f5400e-ce4e-405c-aa8a-9a1751f89fe1" (UID: "a3f5400e-ce4e-405c-aa8a-9a1751f89fe1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.292464 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3f5400e-ce4e-405c-aa8a-9a1751f89fe1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a3f5400e-ce4e-405c-aa8a-9a1751f89fe1" (UID: "a3f5400e-ce4e-405c-aa8a-9a1751f89fe1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.300677 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.301919 4873 scope.go:117] "RemoveContainer" containerID="fef3c34401abd0f5d59cd50b1060ddcaf4cba9139a4f8d95d8ee0118208ed803" Dec 01 09:01:57 crc kubenswrapper[4873]: E1201 09:01:57.302488 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fef3c34401abd0f5d59cd50b1060ddcaf4cba9139a4f8d95d8ee0118208ed803\": container with ID starting with fef3c34401abd0f5d59cd50b1060ddcaf4cba9139a4f8d95d8ee0118208ed803 not found: ID does not exist" containerID="fef3c34401abd0f5d59cd50b1060ddcaf4cba9139a4f8d95d8ee0118208ed803" Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.302523 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fef3c34401abd0f5d59cd50b1060ddcaf4cba9139a4f8d95d8ee0118208ed803"} err="failed to get container status \"fef3c34401abd0f5d59cd50b1060ddcaf4cba9139a4f8d95d8ee0118208ed803\": rpc error: code = NotFound desc = could not find container \"fef3c34401abd0f5d59cd50b1060ddcaf4cba9139a4f8d95d8ee0118208ed803\": container with ID starting with fef3c34401abd0f5d59cd50b1060ddcaf4cba9139a4f8d95d8ee0118208ed803 not found: ID does not exist" Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.341072 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2kzsz\" (UniqueName: \"kubernetes.io/projected/a3f5400e-ce4e-405c-aa8a-9a1751f89fe1-kube-api-access-2kzsz\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.341116 4873 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3f5400e-ce4e-405c-aa8a-9a1751f89fe1-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.341129 4873 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3f5400e-ce4e-405c-aa8a-9a1751f89fe1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.442313 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mf5wt\" (UniqueName: \"kubernetes.io/projected/488ee0a6-2ff5-4d21-807a-775375d48151-kube-api-access-mf5wt\") pod \"488ee0a6-2ff5-4d21-807a-775375d48151\" (UID: \"488ee0a6-2ff5-4d21-807a-775375d48151\") " Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.442444 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/488ee0a6-2ff5-4d21-807a-775375d48151-logs\") pod \"488ee0a6-2ff5-4d21-807a-775375d48151\" (UID: \"488ee0a6-2ff5-4d21-807a-775375d48151\") " Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.442625 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/488ee0a6-2ff5-4d21-807a-775375d48151-combined-ca-bundle\") pod \"488ee0a6-2ff5-4d21-807a-775375d48151\" (UID: \"488ee0a6-2ff5-4d21-807a-775375d48151\") " Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.442694 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/488ee0a6-2ff5-4d21-807a-775375d48151-config-data\") pod \"488ee0a6-2ff5-4d21-807a-775375d48151\" (UID: \"488ee0a6-2ff5-4d21-807a-775375d48151\") " Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.443182 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/488ee0a6-2ff5-4d21-807a-775375d48151-logs" (OuterVolumeSpecName: "logs") pod "488ee0a6-2ff5-4d21-807a-775375d48151" (UID: "488ee0a6-2ff5-4d21-807a-775375d48151"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.443709 4873 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/488ee0a6-2ff5-4d21-807a-775375d48151-logs\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.464377 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/488ee0a6-2ff5-4d21-807a-775375d48151-kube-api-access-mf5wt" (OuterVolumeSpecName: "kube-api-access-mf5wt") pod "488ee0a6-2ff5-4d21-807a-775375d48151" (UID: "488ee0a6-2ff5-4d21-807a-775375d48151"). InnerVolumeSpecName "kube-api-access-mf5wt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.480345 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/488ee0a6-2ff5-4d21-807a-775375d48151-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "488ee0a6-2ff5-4d21-807a-775375d48151" (UID: "488ee0a6-2ff5-4d21-807a-775375d48151"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.481713 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/488ee0a6-2ff5-4d21-807a-775375d48151-config-data" (OuterVolumeSpecName: "config-data") pod "488ee0a6-2ff5-4d21-807a-775375d48151" (UID: "488ee0a6-2ff5-4d21-807a-775375d48151"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.545552 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mf5wt\" (UniqueName: \"kubernetes.io/projected/488ee0a6-2ff5-4d21-807a-775375d48151-kube-api-access-mf5wt\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.545590 4873 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/488ee0a6-2ff5-4d21-807a-775375d48151-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.545600 4873 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/488ee0a6-2ff5-4d21-807a-775375d48151-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.560369 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.573113 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.585748 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 09:01:57 crc kubenswrapper[4873]: E1201 09:01:57.586265 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3f5400e-ce4e-405c-aa8a-9a1751f89fe1" containerName="nova-scheduler-scheduler" Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.586286 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3f5400e-ce4e-405c-aa8a-9a1751f89fe1" containerName="nova-scheduler-scheduler" Dec 01 09:01:57 crc kubenswrapper[4873]: E1201 09:01:57.586303 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="488ee0a6-2ff5-4d21-807a-775375d48151" containerName="nova-api-log" Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.586310 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="488ee0a6-2ff5-4d21-807a-775375d48151" containerName="nova-api-log" Dec 01 09:01:57 crc kubenswrapper[4873]: E1201 09:01:57.586337 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="488ee0a6-2ff5-4d21-807a-775375d48151" containerName="nova-api-api" Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.586343 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="488ee0a6-2ff5-4d21-807a-775375d48151" containerName="nova-api-api" Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.586508 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="488ee0a6-2ff5-4d21-807a-775375d48151" containerName="nova-api-log" Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.586527 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3f5400e-ce4e-405c-aa8a-9a1751f89fe1" containerName="nova-scheduler-scheduler" Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.586553 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="488ee0a6-2ff5-4d21-807a-775375d48151" containerName="nova-api-api" Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.587277 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.589915 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.597547 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.649070 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8dce570f-2670-49a0-a42c-8e12d118f279-config-data\") pod \"nova-scheduler-0\" (UID: \"8dce570f-2670-49a0-a42c-8e12d118f279\") " pod="openstack/nova-scheduler-0" Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.649283 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dce570f-2670-49a0-a42c-8e12d118f279-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8dce570f-2670-49a0-a42c-8e12d118f279\") " pod="openstack/nova-scheduler-0" Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.649353 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5gd2\" (UniqueName: \"kubernetes.io/projected/8dce570f-2670-49a0-a42c-8e12d118f279-kube-api-access-w5gd2\") pod \"nova-scheduler-0\" (UID: \"8dce570f-2670-49a0-a42c-8e12d118f279\") " pod="openstack/nova-scheduler-0" Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.752899 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8dce570f-2670-49a0-a42c-8e12d118f279-config-data\") pod \"nova-scheduler-0\" (UID: \"8dce570f-2670-49a0-a42c-8e12d118f279\") " pod="openstack/nova-scheduler-0" Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.753038 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dce570f-2670-49a0-a42c-8e12d118f279-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8dce570f-2670-49a0-a42c-8e12d118f279\") " pod="openstack/nova-scheduler-0" Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.753074 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5gd2\" (UniqueName: \"kubernetes.io/projected/8dce570f-2670-49a0-a42c-8e12d118f279-kube-api-access-w5gd2\") pod \"nova-scheduler-0\" (UID: \"8dce570f-2670-49a0-a42c-8e12d118f279\") " pod="openstack/nova-scheduler-0" Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.758090 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dce570f-2670-49a0-a42c-8e12d118f279-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8dce570f-2670-49a0-a42c-8e12d118f279\") " pod="openstack/nova-scheduler-0" Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.758127 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8dce570f-2670-49a0-a42c-8e12d118f279-config-data\") pod \"nova-scheduler-0\" (UID: \"8dce570f-2670-49a0-a42c-8e12d118f279\") " pod="openstack/nova-scheduler-0" Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.773306 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5gd2\" (UniqueName: \"kubernetes.io/projected/8dce570f-2670-49a0-a42c-8e12d118f279-kube-api-access-w5gd2\") pod \"nova-scheduler-0\" (UID: \"8dce570f-2670-49a0-a42c-8e12d118f279\") " pod="openstack/nova-scheduler-0" Dec 01 09:01:57 crc kubenswrapper[4873]: I1201 09:01:57.917181 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 09:01:58 crc kubenswrapper[4873]: I1201 09:01:58.250574 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"488ee0a6-2ff5-4d21-807a-775375d48151","Type":"ContainerDied","Data":"ad3632d91f7852372b0b8a474085c4599446547613acf55bb04129beeae5c793"} Dec 01 09:01:58 crc kubenswrapper[4873]: I1201 09:01:58.250635 4873 scope.go:117] "RemoveContainer" containerID="4d11339744726b9c76845472b2f90733ebfd28f6543df724571df4976f39b8ea" Dec 01 09:01:58 crc kubenswrapper[4873]: I1201 09:01:58.250676 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 09:01:58 crc kubenswrapper[4873]: I1201 09:01:58.292470 4873 scope.go:117] "RemoveContainer" containerID="0f587475d84c0bcc0cd7d7cac7c5d1235d8cf68292b2b0c2489bab77afa83ebf" Dec 01 09:01:58 crc kubenswrapper[4873]: I1201 09:01:58.301614 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 09:01:58 crc kubenswrapper[4873]: I1201 09:01:58.314755 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 01 09:01:58 crc kubenswrapper[4873]: I1201 09:01:58.328268 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 01 09:01:58 crc kubenswrapper[4873]: I1201 09:01:58.330803 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 09:01:58 crc kubenswrapper[4873]: I1201 09:01:58.333348 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 01 09:01:58 crc kubenswrapper[4873]: I1201 09:01:58.337487 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 09:01:58 crc kubenswrapper[4873]: I1201 09:01:58.377424 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77xng\" (UniqueName: \"kubernetes.io/projected/80b11b29-b572-4fb2-a32e-7529fbf58637-kube-api-access-77xng\") pod \"nova-api-0\" (UID: \"80b11b29-b572-4fb2-a32e-7529fbf58637\") " pod="openstack/nova-api-0" Dec 01 09:01:58 crc kubenswrapper[4873]: I1201 09:01:58.377516 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80b11b29-b572-4fb2-a32e-7529fbf58637-config-data\") pod \"nova-api-0\" (UID: \"80b11b29-b572-4fb2-a32e-7529fbf58637\") " pod="openstack/nova-api-0" Dec 01 09:01:58 crc kubenswrapper[4873]: I1201 09:01:58.377627 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80b11b29-b572-4fb2-a32e-7529fbf58637-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"80b11b29-b572-4fb2-a32e-7529fbf58637\") " pod="openstack/nova-api-0" Dec 01 09:01:58 crc kubenswrapper[4873]: I1201 09:01:58.377660 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80b11b29-b572-4fb2-a32e-7529fbf58637-logs\") pod \"nova-api-0\" (UID: \"80b11b29-b572-4fb2-a32e-7529fbf58637\") " pod="openstack/nova-api-0" Dec 01 09:01:58 crc kubenswrapper[4873]: I1201 09:01:58.444881 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="488ee0a6-2ff5-4d21-807a-775375d48151" path="/var/lib/kubelet/pods/488ee0a6-2ff5-4d21-807a-775375d48151/volumes" Dec 01 09:01:58 crc kubenswrapper[4873]: I1201 09:01:58.445858 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3f5400e-ce4e-405c-aa8a-9a1751f89fe1" path="/var/lib/kubelet/pods/a3f5400e-ce4e-405c-aa8a-9a1751f89fe1/volumes" Dec 01 09:01:58 crc kubenswrapper[4873]: I1201 09:01:58.447607 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 09:01:58 crc kubenswrapper[4873]: I1201 09:01:58.480312 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77xng\" (UniqueName: \"kubernetes.io/projected/80b11b29-b572-4fb2-a32e-7529fbf58637-kube-api-access-77xng\") pod \"nova-api-0\" (UID: \"80b11b29-b572-4fb2-a32e-7529fbf58637\") " pod="openstack/nova-api-0" Dec 01 09:01:58 crc kubenswrapper[4873]: I1201 09:01:58.480512 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80b11b29-b572-4fb2-a32e-7529fbf58637-config-data\") pod \"nova-api-0\" (UID: \"80b11b29-b572-4fb2-a32e-7529fbf58637\") " pod="openstack/nova-api-0" Dec 01 09:01:58 crc kubenswrapper[4873]: I1201 09:01:58.480673 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80b11b29-b572-4fb2-a32e-7529fbf58637-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"80b11b29-b572-4fb2-a32e-7529fbf58637\") " pod="openstack/nova-api-0" Dec 01 09:01:58 crc kubenswrapper[4873]: I1201 09:01:58.480769 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80b11b29-b572-4fb2-a32e-7529fbf58637-logs\") pod \"nova-api-0\" (UID: \"80b11b29-b572-4fb2-a32e-7529fbf58637\") " pod="openstack/nova-api-0" Dec 01 09:01:58 crc kubenswrapper[4873]: I1201 09:01:58.481398 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80b11b29-b572-4fb2-a32e-7529fbf58637-logs\") pod \"nova-api-0\" (UID: \"80b11b29-b572-4fb2-a32e-7529fbf58637\") " pod="openstack/nova-api-0" Dec 01 09:01:58 crc kubenswrapper[4873]: I1201 09:01:58.485117 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80b11b29-b572-4fb2-a32e-7529fbf58637-config-data\") pod \"nova-api-0\" (UID: \"80b11b29-b572-4fb2-a32e-7529fbf58637\") " pod="openstack/nova-api-0" Dec 01 09:01:58 crc kubenswrapper[4873]: I1201 09:01:58.485300 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80b11b29-b572-4fb2-a32e-7529fbf58637-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"80b11b29-b572-4fb2-a32e-7529fbf58637\") " pod="openstack/nova-api-0" Dec 01 09:01:58 crc kubenswrapper[4873]: I1201 09:01:58.499526 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77xng\" (UniqueName: \"kubernetes.io/projected/80b11b29-b572-4fb2-a32e-7529fbf58637-kube-api-access-77xng\") pod \"nova-api-0\" (UID: \"80b11b29-b572-4fb2-a32e-7529fbf58637\") " pod="openstack/nova-api-0" Dec 01 09:01:58 crc kubenswrapper[4873]: I1201 09:01:58.699269 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 09:01:59 crc kubenswrapper[4873]: I1201 09:01:59.205253 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 09:01:59 crc kubenswrapper[4873]: I1201 09:01:59.263219 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"80b11b29-b572-4fb2-a32e-7529fbf58637","Type":"ContainerStarted","Data":"b1d00334f558ca43a0f491b625035608cff2bdee3d9c57aa25ff901208e54037"} Dec 01 09:01:59 crc kubenswrapper[4873]: I1201 09:01:59.266279 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8dce570f-2670-49a0-a42c-8e12d118f279","Type":"ContainerStarted","Data":"fe777aad3bb04934c08d3f70a2b914656cc6fe44825115f3b18a8b4e1faa1d81"} Dec 01 09:01:59 crc kubenswrapper[4873]: I1201 09:01:59.266374 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8dce570f-2670-49a0-a42c-8e12d118f279","Type":"ContainerStarted","Data":"30d6f7e463dede8b1db5f34b9b6dabbc767e2d4843b3cc782c205f86c3e2db63"} Dec 01 09:01:59 crc kubenswrapper[4873]: I1201 09:01:59.305817 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.305787836 podStartE2EDuration="2.305787836s" podCreationTimestamp="2025-12-01 09:01:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:01:59.299327478 +0000 UTC m=+1295.201436017" watchObservedRunningTime="2025-12-01 09:01:59.305787836 +0000 UTC m=+1295.207896365" Dec 01 09:02:00 crc kubenswrapper[4873]: I1201 09:02:00.282965 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"80b11b29-b572-4fb2-a32e-7529fbf58637","Type":"ContainerStarted","Data":"c7ccfd4c2e96d8e165124420150043cd20cd95accfc258abd83cd6440a508c18"} Dec 01 09:02:00 crc kubenswrapper[4873]: I1201 09:02:00.283655 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"80b11b29-b572-4fb2-a32e-7529fbf58637","Type":"ContainerStarted","Data":"03cb9c7b4a0410bea1842a201e610d35b17a50a977a7b82c45562c3afbf8f908"} Dec 01 09:02:00 crc kubenswrapper[4873]: I1201 09:02:00.306695 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.306663383 podStartE2EDuration="2.306663383s" podCreationTimestamp="2025-12-01 09:01:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:02:00.303352122 +0000 UTC m=+1296.205460691" watchObservedRunningTime="2025-12-01 09:02:00.306663383 +0000 UTC m=+1296.208771952" Dec 01 09:02:01 crc kubenswrapper[4873]: I1201 09:02:01.059140 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:02:01 crc kubenswrapper[4873]: I1201 09:02:01.059609 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:02:02 crc kubenswrapper[4873]: I1201 09:02:02.511050 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 01 09:02:02 crc kubenswrapper[4873]: I1201 09:02:02.918375 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 01 09:02:07 crc kubenswrapper[4873]: I1201 09:02:07.918063 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 01 09:02:07 crc kubenswrapper[4873]: I1201 09:02:07.974223 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 01 09:02:08 crc kubenswrapper[4873]: I1201 09:02:08.444310 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 01 09:02:08 crc kubenswrapper[4873]: I1201 09:02:08.700520 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 09:02:08 crc kubenswrapper[4873]: I1201 09:02:08.700750 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 09:02:09 crc kubenswrapper[4873]: I1201 09:02:09.787331 4873 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="80b11b29-b572-4fb2-a32e-7529fbf58637" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.178:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 01 09:02:09 crc kubenswrapper[4873]: I1201 09:02:09.787386 4873 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="80b11b29-b572-4fb2-a32e-7529fbf58637" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.178:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.372933 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.381736 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.499425 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-68rnr\" (UniqueName: \"kubernetes.io/projected/d271d294-edf0-445a-8df2-7dd767c99d91-kube-api-access-68rnr\") pod \"d271d294-edf0-445a-8df2-7dd767c99d91\" (UID: \"d271d294-edf0-445a-8df2-7dd767c99d91\") " Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.499562 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d271d294-edf0-445a-8df2-7dd767c99d91-config-data\") pod \"d271d294-edf0-445a-8df2-7dd767c99d91\" (UID: \"d271d294-edf0-445a-8df2-7dd767c99d91\") " Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.499711 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d271d294-edf0-445a-8df2-7dd767c99d91-combined-ca-bundle\") pod \"d271d294-edf0-445a-8df2-7dd767c99d91\" (UID: \"d271d294-edf0-445a-8df2-7dd767c99d91\") " Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.499776 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa610495-1013-47ed-b54d-c83962c8caf2-config-data\") pod \"fa610495-1013-47ed-b54d-c83962c8caf2\" (UID: \"fa610495-1013-47ed-b54d-c83962c8caf2\") " Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.499803 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8ln7\" (UniqueName: \"kubernetes.io/projected/fa610495-1013-47ed-b54d-c83962c8caf2-kube-api-access-g8ln7\") pod \"fa610495-1013-47ed-b54d-c83962c8caf2\" (UID: \"fa610495-1013-47ed-b54d-c83962c8caf2\") " Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.499918 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa610495-1013-47ed-b54d-c83962c8caf2-combined-ca-bundle\") pod \"fa610495-1013-47ed-b54d-c83962c8caf2\" (UID: \"fa610495-1013-47ed-b54d-c83962c8caf2\") " Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.509265 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d271d294-edf0-445a-8df2-7dd767c99d91-logs\") pod \"d271d294-edf0-445a-8df2-7dd767c99d91\" (UID: \"d271d294-edf0-445a-8df2-7dd767c99d91\") " Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.511038 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d271d294-edf0-445a-8df2-7dd767c99d91-logs" (OuterVolumeSpecName: "logs") pod "d271d294-edf0-445a-8df2-7dd767c99d91" (UID: "d271d294-edf0-445a-8df2-7dd767c99d91"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.586626 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa610495-1013-47ed-b54d-c83962c8caf2-kube-api-access-g8ln7" (OuterVolumeSpecName: "kube-api-access-g8ln7") pod "fa610495-1013-47ed-b54d-c83962c8caf2" (UID: "fa610495-1013-47ed-b54d-c83962c8caf2"). InnerVolumeSpecName "kube-api-access-g8ln7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.605452 4873 generic.go:334] "Generic (PLEG): container finished" podID="d271d294-edf0-445a-8df2-7dd767c99d91" containerID="5f77e4d0b56d5f8df84215b671d6f830b670a5adfa5fdedb52c448d3c18da079" exitCode=137 Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.605570 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d271d294-edf0-445a-8df2-7dd767c99d91","Type":"ContainerDied","Data":"5f77e4d0b56d5f8df84215b671d6f830b670a5adfa5fdedb52c448d3c18da079"} Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.605612 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d271d294-edf0-445a-8df2-7dd767c99d91","Type":"ContainerDied","Data":"002fe8ab9337f03f01c21f6c2e3004c118d8495bce9a5ea52df893c72990d090"} Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.605633 4873 scope.go:117] "RemoveContainer" containerID="5f77e4d0b56d5f8df84215b671d6f830b670a5adfa5fdedb52c448d3c18da079" Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.605850 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.616248 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d271d294-edf0-445a-8df2-7dd767c99d91-kube-api-access-68rnr" (OuterVolumeSpecName: "kube-api-access-68rnr") pod "d271d294-edf0-445a-8df2-7dd767c99d91" (UID: "d271d294-edf0-445a-8df2-7dd767c99d91"). InnerVolumeSpecName "kube-api-access-68rnr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.618445 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8ln7\" (UniqueName: \"kubernetes.io/projected/fa610495-1013-47ed-b54d-c83962c8caf2-kube-api-access-g8ln7\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.618463 4873 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d271d294-edf0-445a-8df2-7dd767c99d91-logs\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.618475 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-68rnr\" (UniqueName: \"kubernetes.io/projected/d271d294-edf0-445a-8df2-7dd767c99d91-kube-api-access-68rnr\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.646203 4873 generic.go:334] "Generic (PLEG): container finished" podID="fa610495-1013-47ed-b54d-c83962c8caf2" containerID="2100353d6abdda91e019b49423ec64a465c08416e6116500b8f57695fcd32b45" exitCode=137 Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.646365 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.646461 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"fa610495-1013-47ed-b54d-c83962c8caf2","Type":"ContainerDied","Data":"2100353d6abdda91e019b49423ec64a465c08416e6116500b8f57695fcd32b45"} Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.646550 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"fa610495-1013-47ed-b54d-c83962c8caf2","Type":"ContainerDied","Data":"754dafd703cdbe9c70fb79bbfc368cbf1bcf907ab0b2aa0978cc1b36610a975b"} Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.691220 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa610495-1013-47ed-b54d-c83962c8caf2-config-data" (OuterVolumeSpecName: "config-data") pod "fa610495-1013-47ed-b54d-c83962c8caf2" (UID: "fa610495-1013-47ed-b54d-c83962c8caf2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.701540 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa610495-1013-47ed-b54d-c83962c8caf2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fa610495-1013-47ed-b54d-c83962c8caf2" (UID: "fa610495-1013-47ed-b54d-c83962c8caf2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.705378 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d271d294-edf0-445a-8df2-7dd767c99d91-config-data" (OuterVolumeSpecName: "config-data") pod "d271d294-edf0-445a-8df2-7dd767c99d91" (UID: "d271d294-edf0-445a-8df2-7dd767c99d91"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.709723 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d271d294-edf0-445a-8df2-7dd767c99d91-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d271d294-edf0-445a-8df2-7dd767c99d91" (UID: "d271d294-edf0-445a-8df2-7dd767c99d91"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.719913 4873 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d271d294-edf0-445a-8df2-7dd767c99d91-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.719942 4873 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa610495-1013-47ed-b54d-c83962c8caf2-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.719951 4873 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa610495-1013-47ed-b54d-c83962c8caf2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.719960 4873 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d271d294-edf0-445a-8df2-7dd767c99d91-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.727390 4873 scope.go:117] "RemoveContainer" containerID="569632beef6059c5d114c426d47b88d1797622618401a6f265c69e17ae7ed0d0" Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.751709 4873 scope.go:117] "RemoveContainer" containerID="5f77e4d0b56d5f8df84215b671d6f830b670a5adfa5fdedb52c448d3c18da079" Dec 01 09:02:15 crc kubenswrapper[4873]: E1201 09:02:15.752054 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f77e4d0b56d5f8df84215b671d6f830b670a5adfa5fdedb52c448d3c18da079\": container with ID starting with 5f77e4d0b56d5f8df84215b671d6f830b670a5adfa5fdedb52c448d3c18da079 not found: ID does not exist" containerID="5f77e4d0b56d5f8df84215b671d6f830b670a5adfa5fdedb52c448d3c18da079" Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.752087 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f77e4d0b56d5f8df84215b671d6f830b670a5adfa5fdedb52c448d3c18da079"} err="failed to get container status \"5f77e4d0b56d5f8df84215b671d6f830b670a5adfa5fdedb52c448d3c18da079\": rpc error: code = NotFound desc = could not find container \"5f77e4d0b56d5f8df84215b671d6f830b670a5adfa5fdedb52c448d3c18da079\": container with ID starting with 5f77e4d0b56d5f8df84215b671d6f830b670a5adfa5fdedb52c448d3c18da079 not found: ID does not exist" Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.752114 4873 scope.go:117] "RemoveContainer" containerID="569632beef6059c5d114c426d47b88d1797622618401a6f265c69e17ae7ed0d0" Dec 01 09:02:15 crc kubenswrapper[4873]: E1201 09:02:15.752344 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"569632beef6059c5d114c426d47b88d1797622618401a6f265c69e17ae7ed0d0\": container with ID starting with 569632beef6059c5d114c426d47b88d1797622618401a6f265c69e17ae7ed0d0 not found: ID does not exist" containerID="569632beef6059c5d114c426d47b88d1797622618401a6f265c69e17ae7ed0d0" Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.752373 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"569632beef6059c5d114c426d47b88d1797622618401a6f265c69e17ae7ed0d0"} err="failed to get container status \"569632beef6059c5d114c426d47b88d1797622618401a6f265c69e17ae7ed0d0\": rpc error: code = NotFound desc = could not find container \"569632beef6059c5d114c426d47b88d1797622618401a6f265c69e17ae7ed0d0\": container with ID starting with 569632beef6059c5d114c426d47b88d1797622618401a6f265c69e17ae7ed0d0 not found: ID does not exist" Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.752395 4873 scope.go:117] "RemoveContainer" containerID="2100353d6abdda91e019b49423ec64a465c08416e6116500b8f57695fcd32b45" Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.773950 4873 scope.go:117] "RemoveContainer" containerID="2100353d6abdda91e019b49423ec64a465c08416e6116500b8f57695fcd32b45" Dec 01 09:02:15 crc kubenswrapper[4873]: E1201 09:02:15.774577 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2100353d6abdda91e019b49423ec64a465c08416e6116500b8f57695fcd32b45\": container with ID starting with 2100353d6abdda91e019b49423ec64a465c08416e6116500b8f57695fcd32b45 not found: ID does not exist" containerID="2100353d6abdda91e019b49423ec64a465c08416e6116500b8f57695fcd32b45" Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.774615 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2100353d6abdda91e019b49423ec64a465c08416e6116500b8f57695fcd32b45"} err="failed to get container status \"2100353d6abdda91e019b49423ec64a465c08416e6116500b8f57695fcd32b45\": rpc error: code = NotFound desc = could not find container \"2100353d6abdda91e019b49423ec64a465c08416e6116500b8f57695fcd32b45\": container with ID starting with 2100353d6abdda91e019b49423ec64a465c08416e6116500b8f57695fcd32b45 not found: ID does not exist" Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.947210 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.965280 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.986743 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 01 09:02:15 crc kubenswrapper[4873]: E1201 09:02:15.987210 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d271d294-edf0-445a-8df2-7dd767c99d91" containerName="nova-metadata-metadata" Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.987237 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="d271d294-edf0-445a-8df2-7dd767c99d91" containerName="nova-metadata-metadata" Dec 01 09:02:15 crc kubenswrapper[4873]: E1201 09:02:15.987262 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa610495-1013-47ed-b54d-c83962c8caf2" containerName="nova-cell1-novncproxy-novncproxy" Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.987271 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa610495-1013-47ed-b54d-c83962c8caf2" containerName="nova-cell1-novncproxy-novncproxy" Dec 01 09:02:15 crc kubenswrapper[4873]: E1201 09:02:15.987302 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d271d294-edf0-445a-8df2-7dd767c99d91" containerName="nova-metadata-log" Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.987311 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="d271d294-edf0-445a-8df2-7dd767c99d91" containerName="nova-metadata-log" Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.987512 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa610495-1013-47ed-b54d-c83962c8caf2" containerName="nova-cell1-novncproxy-novncproxy" Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.987536 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="d271d294-edf0-445a-8df2-7dd767c99d91" containerName="nova-metadata-metadata" Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.987562 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="d271d294-edf0-445a-8df2-7dd767c99d91" containerName="nova-metadata-log" Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.988769 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.993940 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 01 09:02:15 crc kubenswrapper[4873]: I1201 09:02:15.994000 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 01 09:02:16 crc kubenswrapper[4873]: I1201 09:02:16.018033 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 09:02:16 crc kubenswrapper[4873]: I1201 09:02:16.027372 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/50f5253a-ab03-426c-aaca-a8db33425fef-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"50f5253a-ab03-426c-aaca-a8db33425fef\") " pod="openstack/nova-metadata-0" Dec 01 09:02:16 crc kubenswrapper[4873]: I1201 09:02:16.027470 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7jdj\" (UniqueName: \"kubernetes.io/projected/50f5253a-ab03-426c-aaca-a8db33425fef-kube-api-access-r7jdj\") pod \"nova-metadata-0\" (UID: \"50f5253a-ab03-426c-aaca-a8db33425fef\") " pod="openstack/nova-metadata-0" Dec 01 09:02:16 crc kubenswrapper[4873]: I1201 09:02:16.027520 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50f5253a-ab03-426c-aaca-a8db33425fef-logs\") pod \"nova-metadata-0\" (UID: \"50f5253a-ab03-426c-aaca-a8db33425fef\") " pod="openstack/nova-metadata-0" Dec 01 09:02:16 crc kubenswrapper[4873]: I1201 09:02:16.027594 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50f5253a-ab03-426c-aaca-a8db33425fef-config-data\") pod \"nova-metadata-0\" (UID: \"50f5253a-ab03-426c-aaca-a8db33425fef\") " pod="openstack/nova-metadata-0" Dec 01 09:02:16 crc kubenswrapper[4873]: I1201 09:02:16.027617 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50f5253a-ab03-426c-aaca-a8db33425fef-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"50f5253a-ab03-426c-aaca-a8db33425fef\") " pod="openstack/nova-metadata-0" Dec 01 09:02:16 crc kubenswrapper[4873]: I1201 09:02:16.038084 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 09:02:16 crc kubenswrapper[4873]: I1201 09:02:16.056619 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 09:02:16 crc kubenswrapper[4873]: I1201 09:02:16.072072 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 09:02:16 crc kubenswrapper[4873]: I1201 09:02:16.073724 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 09:02:16 crc kubenswrapper[4873]: I1201 09:02:16.076703 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 01 09:02:16 crc kubenswrapper[4873]: I1201 09:02:16.076820 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 01 09:02:16 crc kubenswrapper[4873]: I1201 09:02:16.076711 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 01 09:02:16 crc kubenswrapper[4873]: I1201 09:02:16.101516 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 09:02:16 crc kubenswrapper[4873]: I1201 09:02:16.128921 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50f5253a-ab03-426c-aaca-a8db33425fef-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"50f5253a-ab03-426c-aaca-a8db33425fef\") " pod="openstack/nova-metadata-0" Dec 01 09:02:16 crc kubenswrapper[4873]: I1201 09:02:16.129047 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/50f5253a-ab03-426c-aaca-a8db33425fef-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"50f5253a-ab03-426c-aaca-a8db33425fef\") " pod="openstack/nova-metadata-0" Dec 01 09:02:16 crc kubenswrapper[4873]: I1201 09:02:16.129111 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7jdj\" (UniqueName: \"kubernetes.io/projected/50f5253a-ab03-426c-aaca-a8db33425fef-kube-api-access-r7jdj\") pod \"nova-metadata-0\" (UID: \"50f5253a-ab03-426c-aaca-a8db33425fef\") " pod="openstack/nova-metadata-0" Dec 01 09:02:16 crc kubenswrapper[4873]: I1201 09:02:16.129155 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50f5253a-ab03-426c-aaca-a8db33425fef-logs\") pod \"nova-metadata-0\" (UID: \"50f5253a-ab03-426c-aaca-a8db33425fef\") " pod="openstack/nova-metadata-0" Dec 01 09:02:16 crc kubenswrapper[4873]: I1201 09:02:16.129210 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50f5253a-ab03-426c-aaca-a8db33425fef-config-data\") pod \"nova-metadata-0\" (UID: \"50f5253a-ab03-426c-aaca-a8db33425fef\") " pod="openstack/nova-metadata-0" Dec 01 09:02:16 crc kubenswrapper[4873]: I1201 09:02:16.130436 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50f5253a-ab03-426c-aaca-a8db33425fef-logs\") pod \"nova-metadata-0\" (UID: \"50f5253a-ab03-426c-aaca-a8db33425fef\") " pod="openstack/nova-metadata-0" Dec 01 09:02:16 crc kubenswrapper[4873]: I1201 09:02:16.133675 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/50f5253a-ab03-426c-aaca-a8db33425fef-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"50f5253a-ab03-426c-aaca-a8db33425fef\") " pod="openstack/nova-metadata-0" Dec 01 09:02:16 crc kubenswrapper[4873]: I1201 09:02:16.133861 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50f5253a-ab03-426c-aaca-a8db33425fef-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"50f5253a-ab03-426c-aaca-a8db33425fef\") " pod="openstack/nova-metadata-0" Dec 01 09:02:16 crc kubenswrapper[4873]: I1201 09:02:16.138287 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50f5253a-ab03-426c-aaca-a8db33425fef-config-data\") pod \"nova-metadata-0\" (UID: \"50f5253a-ab03-426c-aaca-a8db33425fef\") " pod="openstack/nova-metadata-0" Dec 01 09:02:16 crc kubenswrapper[4873]: I1201 09:02:16.148737 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7jdj\" (UniqueName: \"kubernetes.io/projected/50f5253a-ab03-426c-aaca-a8db33425fef-kube-api-access-r7jdj\") pod \"nova-metadata-0\" (UID: \"50f5253a-ab03-426c-aaca-a8db33425fef\") " pod="openstack/nova-metadata-0" Dec 01 09:02:16 crc kubenswrapper[4873]: I1201 09:02:16.231813 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae20a830-d3d8-42fc-9fff-1aabc6ac8139-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ae20a830-d3d8-42fc-9fff-1aabc6ac8139\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 09:02:16 crc kubenswrapper[4873]: I1201 09:02:16.231915 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgbrp\" (UniqueName: \"kubernetes.io/projected/ae20a830-d3d8-42fc-9fff-1aabc6ac8139-kube-api-access-cgbrp\") pod \"nova-cell1-novncproxy-0\" (UID: \"ae20a830-d3d8-42fc-9fff-1aabc6ac8139\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 09:02:16 crc kubenswrapper[4873]: I1201 09:02:16.232612 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae20a830-d3d8-42fc-9fff-1aabc6ac8139-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ae20a830-d3d8-42fc-9fff-1aabc6ac8139\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 09:02:16 crc kubenswrapper[4873]: I1201 09:02:16.232915 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae20a830-d3d8-42fc-9fff-1aabc6ac8139-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ae20a830-d3d8-42fc-9fff-1aabc6ac8139\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 09:02:16 crc kubenswrapper[4873]: I1201 09:02:16.233086 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae20a830-d3d8-42fc-9fff-1aabc6ac8139-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ae20a830-d3d8-42fc-9fff-1aabc6ac8139\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 09:02:16 crc kubenswrapper[4873]: I1201 09:02:16.335739 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae20a830-d3d8-42fc-9fff-1aabc6ac8139-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ae20a830-d3d8-42fc-9fff-1aabc6ac8139\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 09:02:16 crc kubenswrapper[4873]: I1201 09:02:16.335883 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae20a830-d3d8-42fc-9fff-1aabc6ac8139-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ae20a830-d3d8-42fc-9fff-1aabc6ac8139\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 09:02:16 crc kubenswrapper[4873]: I1201 09:02:16.335921 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae20a830-d3d8-42fc-9fff-1aabc6ac8139-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ae20a830-d3d8-42fc-9fff-1aabc6ac8139\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 09:02:16 crc kubenswrapper[4873]: I1201 09:02:16.335973 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae20a830-d3d8-42fc-9fff-1aabc6ac8139-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ae20a830-d3d8-42fc-9fff-1aabc6ac8139\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 09:02:16 crc kubenswrapper[4873]: I1201 09:02:16.336045 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgbrp\" (UniqueName: \"kubernetes.io/projected/ae20a830-d3d8-42fc-9fff-1aabc6ac8139-kube-api-access-cgbrp\") pod \"nova-cell1-novncproxy-0\" (UID: \"ae20a830-d3d8-42fc-9fff-1aabc6ac8139\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 09:02:16 crc kubenswrapper[4873]: I1201 09:02:16.340668 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae20a830-d3d8-42fc-9fff-1aabc6ac8139-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ae20a830-d3d8-42fc-9fff-1aabc6ac8139\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 09:02:16 crc kubenswrapper[4873]: I1201 09:02:16.340708 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae20a830-d3d8-42fc-9fff-1aabc6ac8139-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ae20a830-d3d8-42fc-9fff-1aabc6ac8139\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 09:02:16 crc kubenswrapper[4873]: I1201 09:02:16.341300 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae20a830-d3d8-42fc-9fff-1aabc6ac8139-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ae20a830-d3d8-42fc-9fff-1aabc6ac8139\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 09:02:16 crc kubenswrapper[4873]: I1201 09:02:16.342655 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae20a830-d3d8-42fc-9fff-1aabc6ac8139-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ae20a830-d3d8-42fc-9fff-1aabc6ac8139\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 09:02:16 crc kubenswrapper[4873]: I1201 09:02:16.359234 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgbrp\" (UniqueName: \"kubernetes.io/projected/ae20a830-d3d8-42fc-9fff-1aabc6ac8139-kube-api-access-cgbrp\") pod \"nova-cell1-novncproxy-0\" (UID: \"ae20a830-d3d8-42fc-9fff-1aabc6ac8139\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 09:02:16 crc kubenswrapper[4873]: I1201 09:02:16.368457 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 09:02:16 crc kubenswrapper[4873]: I1201 09:02:16.394772 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 09:02:16 crc kubenswrapper[4873]: I1201 09:02:16.445768 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d271d294-edf0-445a-8df2-7dd767c99d91" path="/var/lib/kubelet/pods/d271d294-edf0-445a-8df2-7dd767c99d91/volumes" Dec 01 09:02:16 crc kubenswrapper[4873]: I1201 09:02:16.446677 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa610495-1013-47ed-b54d-c83962c8caf2" path="/var/lib/kubelet/pods/fa610495-1013-47ed-b54d-c83962c8caf2/volumes" Dec 01 09:02:16 crc kubenswrapper[4873]: I1201 09:02:16.792101 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 09:02:16 crc kubenswrapper[4873]: W1201 09:02:16.794903 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae20a830_d3d8_42fc_9fff_1aabc6ac8139.slice/crio-f7568ab2a7e0aebf0e87182bca1ad178f0c5496e1d8db15f1b314f124e19e85b WatchSource:0}: Error finding container f7568ab2a7e0aebf0e87182bca1ad178f0c5496e1d8db15f1b314f124e19e85b: Status 404 returned error can't find the container with id f7568ab2a7e0aebf0e87182bca1ad178f0c5496e1d8db15f1b314f124e19e85b Dec 01 09:02:16 crc kubenswrapper[4873]: I1201 09:02:16.878891 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 09:02:16 crc kubenswrapper[4873]: W1201 09:02:16.885751 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod50f5253a_ab03_426c_aaca_a8db33425fef.slice/crio-14ef3625991cc6aa6b15c1ec6d4d095b2b4d41d5d13793f0331f1e64076e5239 WatchSource:0}: Error finding container 14ef3625991cc6aa6b15c1ec6d4d095b2b4d41d5d13793f0331f1e64076e5239: Status 404 returned error can't find the container with id 14ef3625991cc6aa6b15c1ec6d4d095b2b4d41d5d13793f0331f1e64076e5239 Dec 01 09:02:17 crc kubenswrapper[4873]: I1201 09:02:17.679301 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ae20a830-d3d8-42fc-9fff-1aabc6ac8139","Type":"ContainerStarted","Data":"d7a90d24026bfbc1f8b695de8f4289ecda83687832832d381e2c3d7e0230ab57"} Dec 01 09:02:17 crc kubenswrapper[4873]: I1201 09:02:17.679901 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ae20a830-d3d8-42fc-9fff-1aabc6ac8139","Type":"ContainerStarted","Data":"f7568ab2a7e0aebf0e87182bca1ad178f0c5496e1d8db15f1b314f124e19e85b"} Dec 01 09:02:17 crc kubenswrapper[4873]: I1201 09:02:17.682662 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"50f5253a-ab03-426c-aaca-a8db33425fef","Type":"ContainerStarted","Data":"50276f38fa8798d2887c903bee97a511cc88f6de965d5963d74b1f8e57edb89e"} Dec 01 09:02:17 crc kubenswrapper[4873]: I1201 09:02:17.682715 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"50f5253a-ab03-426c-aaca-a8db33425fef","Type":"ContainerStarted","Data":"b962fe9f343506d6931f2e98b12c658595cee840435c60e66d8a85e3351d0dd4"} Dec 01 09:02:17 crc kubenswrapper[4873]: I1201 09:02:17.682728 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"50f5253a-ab03-426c-aaca-a8db33425fef","Type":"ContainerStarted","Data":"14ef3625991cc6aa6b15c1ec6d4d095b2b4d41d5d13793f0331f1e64076e5239"} Dec 01 09:02:17 crc kubenswrapper[4873]: I1201 09:02:17.703267 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=1.703238649 podStartE2EDuration="1.703238649s" podCreationTimestamp="2025-12-01 09:02:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:02:17.700343129 +0000 UTC m=+1313.602451668" watchObservedRunningTime="2025-12-01 09:02:17.703238649 +0000 UTC m=+1313.605347228" Dec 01 09:02:17 crc kubenswrapper[4873]: I1201 09:02:17.736134 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.736098393 podStartE2EDuration="2.736098393s" podCreationTimestamp="2025-12-01 09:02:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:02:17.723391322 +0000 UTC m=+1313.625499861" watchObservedRunningTime="2025-12-01 09:02:17.736098393 +0000 UTC m=+1313.638206942" Dec 01 09:02:18 crc kubenswrapper[4873]: I1201 09:02:18.414975 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 01 09:02:18 crc kubenswrapper[4873]: I1201 09:02:18.704512 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 01 09:02:18 crc kubenswrapper[4873]: I1201 09:02:18.705683 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 01 09:02:18 crc kubenswrapper[4873]: I1201 09:02:18.705784 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 01 09:02:18 crc kubenswrapper[4873]: I1201 09:02:18.710163 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 01 09:02:19 crc kubenswrapper[4873]: I1201 09:02:19.744032 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 01 09:02:19 crc kubenswrapper[4873]: I1201 09:02:19.750570 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 01 09:02:20 crc kubenswrapper[4873]: I1201 09:02:20.028087 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b856c5697-rw4vp"] Dec 01 09:02:20 crc kubenswrapper[4873]: I1201 09:02:20.035227 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b856c5697-rw4vp" Dec 01 09:02:20 crc kubenswrapper[4873]: I1201 09:02:20.047216 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e825def4-1364-4a5b-8896-b8b633a8142c-ovsdbserver-nb\") pod \"dnsmasq-dns-5b856c5697-rw4vp\" (UID: \"e825def4-1364-4a5b-8896-b8b633a8142c\") " pod="openstack/dnsmasq-dns-5b856c5697-rw4vp" Dec 01 09:02:20 crc kubenswrapper[4873]: I1201 09:02:20.047353 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85n65\" (UniqueName: \"kubernetes.io/projected/e825def4-1364-4a5b-8896-b8b633a8142c-kube-api-access-85n65\") pod \"dnsmasq-dns-5b856c5697-rw4vp\" (UID: \"e825def4-1364-4a5b-8896-b8b633a8142c\") " pod="openstack/dnsmasq-dns-5b856c5697-rw4vp" Dec 01 09:02:20 crc kubenswrapper[4873]: I1201 09:02:20.047384 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e825def4-1364-4a5b-8896-b8b633a8142c-ovsdbserver-sb\") pod \"dnsmasq-dns-5b856c5697-rw4vp\" (UID: \"e825def4-1364-4a5b-8896-b8b633a8142c\") " pod="openstack/dnsmasq-dns-5b856c5697-rw4vp" Dec 01 09:02:20 crc kubenswrapper[4873]: I1201 09:02:20.047445 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e825def4-1364-4a5b-8896-b8b633a8142c-config\") pod \"dnsmasq-dns-5b856c5697-rw4vp\" (UID: \"e825def4-1364-4a5b-8896-b8b633a8142c\") " pod="openstack/dnsmasq-dns-5b856c5697-rw4vp" Dec 01 09:02:20 crc kubenswrapper[4873]: I1201 09:02:20.047484 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e825def4-1364-4a5b-8896-b8b633a8142c-dns-svc\") pod \"dnsmasq-dns-5b856c5697-rw4vp\" (UID: \"e825def4-1364-4a5b-8896-b8b633a8142c\") " pod="openstack/dnsmasq-dns-5b856c5697-rw4vp" Dec 01 09:02:20 crc kubenswrapper[4873]: I1201 09:02:20.075846 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b856c5697-rw4vp"] Dec 01 09:02:20 crc kubenswrapper[4873]: I1201 09:02:20.150840 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85n65\" (UniqueName: \"kubernetes.io/projected/e825def4-1364-4a5b-8896-b8b633a8142c-kube-api-access-85n65\") pod \"dnsmasq-dns-5b856c5697-rw4vp\" (UID: \"e825def4-1364-4a5b-8896-b8b633a8142c\") " pod="openstack/dnsmasq-dns-5b856c5697-rw4vp" Dec 01 09:02:20 crc kubenswrapper[4873]: I1201 09:02:20.150922 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e825def4-1364-4a5b-8896-b8b633a8142c-ovsdbserver-sb\") pod \"dnsmasq-dns-5b856c5697-rw4vp\" (UID: \"e825def4-1364-4a5b-8896-b8b633a8142c\") " pod="openstack/dnsmasq-dns-5b856c5697-rw4vp" Dec 01 09:02:20 crc kubenswrapper[4873]: I1201 09:02:20.150992 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e825def4-1364-4a5b-8896-b8b633a8142c-config\") pod \"dnsmasq-dns-5b856c5697-rw4vp\" (UID: \"e825def4-1364-4a5b-8896-b8b633a8142c\") " pod="openstack/dnsmasq-dns-5b856c5697-rw4vp" Dec 01 09:02:20 crc kubenswrapper[4873]: I1201 09:02:20.151059 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e825def4-1364-4a5b-8896-b8b633a8142c-dns-svc\") pod \"dnsmasq-dns-5b856c5697-rw4vp\" (UID: \"e825def4-1364-4a5b-8896-b8b633a8142c\") " pod="openstack/dnsmasq-dns-5b856c5697-rw4vp" Dec 01 09:02:20 crc kubenswrapper[4873]: I1201 09:02:20.151119 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e825def4-1364-4a5b-8896-b8b633a8142c-ovsdbserver-nb\") pod \"dnsmasq-dns-5b856c5697-rw4vp\" (UID: \"e825def4-1364-4a5b-8896-b8b633a8142c\") " pod="openstack/dnsmasq-dns-5b856c5697-rw4vp" Dec 01 09:02:20 crc kubenswrapper[4873]: I1201 09:02:20.155782 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e825def4-1364-4a5b-8896-b8b633a8142c-config\") pod \"dnsmasq-dns-5b856c5697-rw4vp\" (UID: \"e825def4-1364-4a5b-8896-b8b633a8142c\") " pod="openstack/dnsmasq-dns-5b856c5697-rw4vp" Dec 01 09:02:20 crc kubenswrapper[4873]: I1201 09:02:20.155790 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e825def4-1364-4a5b-8896-b8b633a8142c-dns-svc\") pod \"dnsmasq-dns-5b856c5697-rw4vp\" (UID: \"e825def4-1364-4a5b-8896-b8b633a8142c\") " pod="openstack/dnsmasq-dns-5b856c5697-rw4vp" Dec 01 09:02:20 crc kubenswrapper[4873]: I1201 09:02:20.158062 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e825def4-1364-4a5b-8896-b8b633a8142c-ovsdbserver-sb\") pod \"dnsmasq-dns-5b856c5697-rw4vp\" (UID: \"e825def4-1364-4a5b-8896-b8b633a8142c\") " pod="openstack/dnsmasq-dns-5b856c5697-rw4vp" Dec 01 09:02:20 crc kubenswrapper[4873]: I1201 09:02:20.160234 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e825def4-1364-4a5b-8896-b8b633a8142c-ovsdbserver-nb\") pod \"dnsmasq-dns-5b856c5697-rw4vp\" (UID: \"e825def4-1364-4a5b-8896-b8b633a8142c\") " pod="openstack/dnsmasq-dns-5b856c5697-rw4vp" Dec 01 09:02:20 crc kubenswrapper[4873]: I1201 09:02:20.190165 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85n65\" (UniqueName: \"kubernetes.io/projected/e825def4-1364-4a5b-8896-b8b633a8142c-kube-api-access-85n65\") pod \"dnsmasq-dns-5b856c5697-rw4vp\" (UID: \"e825def4-1364-4a5b-8896-b8b633a8142c\") " pod="openstack/dnsmasq-dns-5b856c5697-rw4vp" Dec 01 09:02:20 crc kubenswrapper[4873]: I1201 09:02:20.358581 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b856c5697-rw4vp" Dec 01 09:02:20 crc kubenswrapper[4873]: I1201 09:02:20.896489 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b856c5697-rw4vp"] Dec 01 09:02:21 crc kubenswrapper[4873]: I1201 09:02:21.368656 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 01 09:02:21 crc kubenswrapper[4873]: I1201 09:02:21.369292 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 01 09:02:21 crc kubenswrapper[4873]: I1201 09:02:21.395217 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 01 09:02:21 crc kubenswrapper[4873]: I1201 09:02:21.764509 4873 generic.go:334] "Generic (PLEG): container finished" podID="e825def4-1364-4a5b-8896-b8b633a8142c" containerID="d658bdfceb80bae9dd70080cb1de56f99de4162da33c63f48de1866ad682e7e7" exitCode=0 Dec 01 09:02:21 crc kubenswrapper[4873]: I1201 09:02:21.764637 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b856c5697-rw4vp" event={"ID":"e825def4-1364-4a5b-8896-b8b633a8142c","Type":"ContainerDied","Data":"d658bdfceb80bae9dd70080cb1de56f99de4162da33c63f48de1866ad682e7e7"} Dec 01 09:02:21 crc kubenswrapper[4873]: I1201 09:02:21.765368 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b856c5697-rw4vp" event={"ID":"e825def4-1364-4a5b-8896-b8b633a8142c","Type":"ContainerStarted","Data":"80f2df72089fa7908fc4b20d86ac3b58d77f85782af61ece0e62922200286d8d"} Dec 01 09:02:22 crc kubenswrapper[4873]: I1201 09:02:22.415313 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 09:02:22 crc kubenswrapper[4873]: I1201 09:02:22.417206 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2d169f6e-9eff-40ba-bd22-d38a10f48bd5" containerName="ceilometer-central-agent" containerID="cri-o://482dc025849cea6c027c087f224133403794cc1bc59ebf2b363714bff1f7756a" gracePeriod=30 Dec 01 09:02:22 crc kubenswrapper[4873]: I1201 09:02:22.417440 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2d169f6e-9eff-40ba-bd22-d38a10f48bd5" containerName="ceilometer-notification-agent" containerID="cri-o://e77a90d91634036934dd350f3ef20bf08c149543c65b347a1517af80f56cabc4" gracePeriod=30 Dec 01 09:02:22 crc kubenswrapper[4873]: I1201 09:02:22.417486 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2d169f6e-9eff-40ba-bd22-d38a10f48bd5" containerName="sg-core" containerID="cri-o://0bf51c0004c4ed88f9ae017b9c36b6bb3c3ffd1058b297d7eea4ff97bbe9653f" gracePeriod=30 Dec 01 09:02:22 crc kubenswrapper[4873]: I1201 09:02:22.417815 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2d169f6e-9eff-40ba-bd22-d38a10f48bd5" containerName="proxy-httpd" containerID="cri-o://809dd553e4c28663ea69e46f77c04be70e3a1947c263b42d2f0e2bae380edb10" gracePeriod=30 Dec 01 09:02:22 crc kubenswrapper[4873]: I1201 09:02:22.759628 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 09:02:22 crc kubenswrapper[4873]: I1201 09:02:22.778286 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b856c5697-rw4vp" event={"ID":"e825def4-1364-4a5b-8896-b8b633a8142c","Type":"ContainerStarted","Data":"05c253da90588e1e95a9fed1a64c169ae4c1cfbcbeaf1410c5a179ba9d783548"} Dec 01 09:02:22 crc kubenswrapper[4873]: I1201 09:02:22.778867 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b856c5697-rw4vp" Dec 01 09:02:22 crc kubenswrapper[4873]: I1201 09:02:22.782336 4873 generic.go:334] "Generic (PLEG): container finished" podID="2d169f6e-9eff-40ba-bd22-d38a10f48bd5" containerID="809dd553e4c28663ea69e46f77c04be70e3a1947c263b42d2f0e2bae380edb10" exitCode=0 Dec 01 09:02:22 crc kubenswrapper[4873]: I1201 09:02:22.782382 4873 generic.go:334] "Generic (PLEG): container finished" podID="2d169f6e-9eff-40ba-bd22-d38a10f48bd5" containerID="0bf51c0004c4ed88f9ae017b9c36b6bb3c3ffd1058b297d7eea4ff97bbe9653f" exitCode=2 Dec 01 09:02:22 crc kubenswrapper[4873]: I1201 09:02:22.782575 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d169f6e-9eff-40ba-bd22-d38a10f48bd5","Type":"ContainerDied","Data":"809dd553e4c28663ea69e46f77c04be70e3a1947c263b42d2f0e2bae380edb10"} Dec 01 09:02:22 crc kubenswrapper[4873]: I1201 09:02:22.782779 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d169f6e-9eff-40ba-bd22-d38a10f48bd5","Type":"ContainerDied","Data":"0bf51c0004c4ed88f9ae017b9c36b6bb3c3ffd1058b297d7eea4ff97bbe9653f"} Dec 01 09:02:22 crc kubenswrapper[4873]: I1201 09:02:22.782650 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="80b11b29-b572-4fb2-a32e-7529fbf58637" containerName="nova-api-log" containerID="cri-o://03cb9c7b4a0410bea1842a201e610d35b17a50a977a7b82c45562c3afbf8f908" gracePeriod=30 Dec 01 09:02:22 crc kubenswrapper[4873]: I1201 09:02:22.782718 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="80b11b29-b572-4fb2-a32e-7529fbf58637" containerName="nova-api-api" containerID="cri-o://c7ccfd4c2e96d8e165124420150043cd20cd95accfc258abd83cd6440a508c18" gracePeriod=30 Dec 01 09:02:22 crc kubenswrapper[4873]: I1201 09:02:22.813279 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b856c5697-rw4vp" podStartSLOduration=3.813251101 podStartE2EDuration="3.813251101s" podCreationTimestamp="2025-12-01 09:02:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:02:22.799292339 +0000 UTC m=+1318.701400878" watchObservedRunningTime="2025-12-01 09:02:22.813251101 +0000 UTC m=+1318.715359640" Dec 01 09:02:23 crc kubenswrapper[4873]: I1201 09:02:23.803460 4873 generic.go:334] "Generic (PLEG): container finished" podID="2d169f6e-9eff-40ba-bd22-d38a10f48bd5" containerID="e77a90d91634036934dd350f3ef20bf08c149543c65b347a1517af80f56cabc4" exitCode=0 Dec 01 09:02:23 crc kubenswrapper[4873]: I1201 09:02:23.803896 4873 generic.go:334] "Generic (PLEG): container finished" podID="2d169f6e-9eff-40ba-bd22-d38a10f48bd5" containerID="482dc025849cea6c027c087f224133403794cc1bc59ebf2b363714bff1f7756a" exitCode=0 Dec 01 09:02:23 crc kubenswrapper[4873]: I1201 09:02:23.803611 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d169f6e-9eff-40ba-bd22-d38a10f48bd5","Type":"ContainerDied","Data":"e77a90d91634036934dd350f3ef20bf08c149543c65b347a1517af80f56cabc4"} Dec 01 09:02:23 crc kubenswrapper[4873]: I1201 09:02:23.804062 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d169f6e-9eff-40ba-bd22-d38a10f48bd5","Type":"ContainerDied","Data":"482dc025849cea6c027c087f224133403794cc1bc59ebf2b363714bff1f7756a"} Dec 01 09:02:23 crc kubenswrapper[4873]: I1201 09:02:23.810706 4873 generic.go:334] "Generic (PLEG): container finished" podID="80b11b29-b572-4fb2-a32e-7529fbf58637" containerID="03cb9c7b4a0410bea1842a201e610d35b17a50a977a7b82c45562c3afbf8f908" exitCode=143 Dec 01 09:02:23 crc kubenswrapper[4873]: I1201 09:02:23.811063 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"80b11b29-b572-4fb2-a32e-7529fbf58637","Type":"ContainerDied","Data":"03cb9c7b4a0410bea1842a201e610d35b17a50a977a7b82c45562c3afbf8f908"} Dec 01 09:02:24 crc kubenswrapper[4873]: I1201 09:02:24.153427 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 09:02:24 crc kubenswrapper[4873]: I1201 09:02:24.352683 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-sg-core-conf-yaml\") pod \"2d169f6e-9eff-40ba-bd22-d38a10f48bd5\" (UID: \"2d169f6e-9eff-40ba-bd22-d38a10f48bd5\") " Dec 01 09:02:24 crc kubenswrapper[4873]: I1201 09:02:24.352813 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-ceilometer-tls-certs\") pod \"2d169f6e-9eff-40ba-bd22-d38a10f48bd5\" (UID: \"2d169f6e-9eff-40ba-bd22-d38a10f48bd5\") " Dec 01 09:02:24 crc kubenswrapper[4873]: I1201 09:02:24.352917 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-config-data\") pod \"2d169f6e-9eff-40ba-bd22-d38a10f48bd5\" (UID: \"2d169f6e-9eff-40ba-bd22-d38a10f48bd5\") " Dec 01 09:02:24 crc kubenswrapper[4873]: I1201 09:02:24.352998 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-scripts\") pod \"2d169f6e-9eff-40ba-bd22-d38a10f48bd5\" (UID: \"2d169f6e-9eff-40ba-bd22-d38a10f48bd5\") " Dec 01 09:02:24 crc kubenswrapper[4873]: I1201 09:02:24.353076 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2hvx\" (UniqueName: \"kubernetes.io/projected/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-kube-api-access-q2hvx\") pod \"2d169f6e-9eff-40ba-bd22-d38a10f48bd5\" (UID: \"2d169f6e-9eff-40ba-bd22-d38a10f48bd5\") " Dec 01 09:02:24 crc kubenswrapper[4873]: I1201 09:02:24.353195 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-combined-ca-bundle\") pod \"2d169f6e-9eff-40ba-bd22-d38a10f48bd5\" (UID: \"2d169f6e-9eff-40ba-bd22-d38a10f48bd5\") " Dec 01 09:02:24 crc kubenswrapper[4873]: I1201 09:02:24.353265 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-run-httpd\") pod \"2d169f6e-9eff-40ba-bd22-d38a10f48bd5\" (UID: \"2d169f6e-9eff-40ba-bd22-d38a10f48bd5\") " Dec 01 09:02:24 crc kubenswrapper[4873]: I1201 09:02:24.353357 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-log-httpd\") pod \"2d169f6e-9eff-40ba-bd22-d38a10f48bd5\" (UID: \"2d169f6e-9eff-40ba-bd22-d38a10f48bd5\") " Dec 01 09:02:24 crc kubenswrapper[4873]: I1201 09:02:24.353885 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2d169f6e-9eff-40ba-bd22-d38a10f48bd5" (UID: "2d169f6e-9eff-40ba-bd22-d38a10f48bd5"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:02:24 crc kubenswrapper[4873]: I1201 09:02:24.354000 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2d169f6e-9eff-40ba-bd22-d38a10f48bd5" (UID: "2d169f6e-9eff-40ba-bd22-d38a10f48bd5"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:02:24 crc kubenswrapper[4873]: I1201 09:02:24.354894 4873 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:24 crc kubenswrapper[4873]: I1201 09:02:24.355152 4873 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:24 crc kubenswrapper[4873]: I1201 09:02:24.368563 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-scripts" (OuterVolumeSpecName: "scripts") pod "2d169f6e-9eff-40ba-bd22-d38a10f48bd5" (UID: "2d169f6e-9eff-40ba-bd22-d38a10f48bd5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:02:24 crc kubenswrapper[4873]: I1201 09:02:24.369489 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-kube-api-access-q2hvx" (OuterVolumeSpecName: "kube-api-access-q2hvx") pod "2d169f6e-9eff-40ba-bd22-d38a10f48bd5" (UID: "2d169f6e-9eff-40ba-bd22-d38a10f48bd5"). InnerVolumeSpecName "kube-api-access-q2hvx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:02:24 crc kubenswrapper[4873]: I1201 09:02:24.396936 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2d169f6e-9eff-40ba-bd22-d38a10f48bd5" (UID: "2d169f6e-9eff-40ba-bd22-d38a10f48bd5"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:02:24 crc kubenswrapper[4873]: I1201 09:02:24.428297 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "2d169f6e-9eff-40ba-bd22-d38a10f48bd5" (UID: "2d169f6e-9eff-40ba-bd22-d38a10f48bd5"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:02:24 crc kubenswrapper[4873]: I1201 09:02:24.456679 4873 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:24 crc kubenswrapper[4873]: I1201 09:02:24.459892 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2hvx\" (UniqueName: \"kubernetes.io/projected/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-kube-api-access-q2hvx\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:24 crc kubenswrapper[4873]: I1201 09:02:24.460074 4873 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:24 crc kubenswrapper[4873]: I1201 09:02:24.460173 4873 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:24 crc kubenswrapper[4873]: I1201 09:02:24.472965 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2d169f6e-9eff-40ba-bd22-d38a10f48bd5" (UID: "2d169f6e-9eff-40ba-bd22-d38a10f48bd5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:02:24 crc kubenswrapper[4873]: I1201 09:02:24.494810 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-config-data" (OuterVolumeSpecName: "config-data") pod "2d169f6e-9eff-40ba-bd22-d38a10f48bd5" (UID: "2d169f6e-9eff-40ba-bd22-d38a10f48bd5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:02:24 crc kubenswrapper[4873]: I1201 09:02:24.563148 4873 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:24 crc kubenswrapper[4873]: I1201 09:02:24.563185 4873 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d169f6e-9eff-40ba-bd22-d38a10f48bd5-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:24 crc kubenswrapper[4873]: I1201 09:02:24.831823 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d169f6e-9eff-40ba-bd22-d38a10f48bd5","Type":"ContainerDied","Data":"8b8a768ed07ad79ec2753c8f1149d6273b521a25aa03a54180667bc61c945aec"} Dec 01 09:02:24 crc kubenswrapper[4873]: I1201 09:02:24.831986 4873 scope.go:117] "RemoveContainer" containerID="809dd553e4c28663ea69e46f77c04be70e3a1947c263b42d2f0e2bae380edb10" Dec 01 09:02:24 crc kubenswrapper[4873]: I1201 09:02:24.831875 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 09:02:24 crc kubenswrapper[4873]: I1201 09:02:24.860115 4873 scope.go:117] "RemoveContainer" containerID="0bf51c0004c4ed88f9ae017b9c36b6bb3c3ffd1058b297d7eea4ff97bbe9653f" Dec 01 09:02:24 crc kubenswrapper[4873]: I1201 09:02:24.893415 4873 scope.go:117] "RemoveContainer" containerID="e77a90d91634036934dd350f3ef20bf08c149543c65b347a1517af80f56cabc4" Dec 01 09:02:24 crc kubenswrapper[4873]: I1201 09:02:24.936182 4873 scope.go:117] "RemoveContainer" containerID="482dc025849cea6c027c087f224133403794cc1bc59ebf2b363714bff1f7756a" Dec 01 09:02:25 crc kubenswrapper[4873]: I1201 09:02:25.011260 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 09:02:25 crc kubenswrapper[4873]: I1201 09:02:25.022236 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 09:02:25 crc kubenswrapper[4873]: I1201 09:02:25.046441 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 09:02:25 crc kubenswrapper[4873]: E1201 09:02:25.047124 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d169f6e-9eff-40ba-bd22-d38a10f48bd5" containerName="ceilometer-notification-agent" Dec 01 09:02:25 crc kubenswrapper[4873]: I1201 09:02:25.047150 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d169f6e-9eff-40ba-bd22-d38a10f48bd5" containerName="ceilometer-notification-agent" Dec 01 09:02:25 crc kubenswrapper[4873]: E1201 09:02:25.047203 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d169f6e-9eff-40ba-bd22-d38a10f48bd5" containerName="sg-core" Dec 01 09:02:25 crc kubenswrapper[4873]: I1201 09:02:25.047211 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d169f6e-9eff-40ba-bd22-d38a10f48bd5" containerName="sg-core" Dec 01 09:02:25 crc kubenswrapper[4873]: E1201 09:02:25.047227 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d169f6e-9eff-40ba-bd22-d38a10f48bd5" containerName="ceilometer-central-agent" Dec 01 09:02:25 crc kubenswrapper[4873]: I1201 09:02:25.047236 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d169f6e-9eff-40ba-bd22-d38a10f48bd5" containerName="ceilometer-central-agent" Dec 01 09:02:25 crc kubenswrapper[4873]: E1201 09:02:25.047244 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d169f6e-9eff-40ba-bd22-d38a10f48bd5" containerName="proxy-httpd" Dec 01 09:02:25 crc kubenswrapper[4873]: I1201 09:02:25.047250 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d169f6e-9eff-40ba-bd22-d38a10f48bd5" containerName="proxy-httpd" Dec 01 09:02:25 crc kubenswrapper[4873]: I1201 09:02:25.047428 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d169f6e-9eff-40ba-bd22-d38a10f48bd5" containerName="ceilometer-notification-agent" Dec 01 09:02:25 crc kubenswrapper[4873]: I1201 09:02:25.047444 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d169f6e-9eff-40ba-bd22-d38a10f48bd5" containerName="sg-core" Dec 01 09:02:25 crc kubenswrapper[4873]: I1201 09:02:25.047459 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d169f6e-9eff-40ba-bd22-d38a10f48bd5" containerName="ceilometer-central-agent" Dec 01 09:02:25 crc kubenswrapper[4873]: I1201 09:02:25.047471 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d169f6e-9eff-40ba-bd22-d38a10f48bd5" containerName="proxy-httpd" Dec 01 09:02:25 crc kubenswrapper[4873]: I1201 09:02:25.049795 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 09:02:25 crc kubenswrapper[4873]: I1201 09:02:25.059789 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 09:02:25 crc kubenswrapper[4873]: I1201 09:02:25.059959 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 01 09:02:25 crc kubenswrapper[4873]: I1201 09:02:25.062932 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 09:02:25 crc kubenswrapper[4873]: I1201 09:02:25.074654 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 09:02:25 crc kubenswrapper[4873]: I1201 09:02:25.101184 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7srnd\" (UniqueName: \"kubernetes.io/projected/c7cad829-b344-4e49-b0d4-264750c0ae0f-kube-api-access-7srnd\") pod \"ceilometer-0\" (UID: \"c7cad829-b344-4e49-b0d4-264750c0ae0f\") " pod="openstack/ceilometer-0" Dec 01 09:02:25 crc kubenswrapper[4873]: I1201 09:02:25.101239 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7cad829-b344-4e49-b0d4-264750c0ae0f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c7cad829-b344-4e49-b0d4-264750c0ae0f\") " pod="openstack/ceilometer-0" Dec 01 09:02:25 crc kubenswrapper[4873]: I1201 09:02:25.101268 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7cad829-b344-4e49-b0d4-264750c0ae0f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c7cad829-b344-4e49-b0d4-264750c0ae0f\") " pod="openstack/ceilometer-0" Dec 01 09:02:25 crc kubenswrapper[4873]: I1201 09:02:25.101309 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c7cad829-b344-4e49-b0d4-264750c0ae0f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c7cad829-b344-4e49-b0d4-264750c0ae0f\") " pod="openstack/ceilometer-0" Dec 01 09:02:25 crc kubenswrapper[4873]: I1201 09:02:25.101336 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7cad829-b344-4e49-b0d4-264750c0ae0f-config-data\") pod \"ceilometer-0\" (UID: \"c7cad829-b344-4e49-b0d4-264750c0ae0f\") " pod="openstack/ceilometer-0" Dec 01 09:02:25 crc kubenswrapper[4873]: I1201 09:02:25.101382 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7cad829-b344-4e49-b0d4-264750c0ae0f-run-httpd\") pod \"ceilometer-0\" (UID: \"c7cad829-b344-4e49-b0d4-264750c0ae0f\") " pod="openstack/ceilometer-0" Dec 01 09:02:25 crc kubenswrapper[4873]: I1201 09:02:25.101397 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7cad829-b344-4e49-b0d4-264750c0ae0f-scripts\") pod \"ceilometer-0\" (UID: \"c7cad829-b344-4e49-b0d4-264750c0ae0f\") " pod="openstack/ceilometer-0" Dec 01 09:02:25 crc kubenswrapper[4873]: I1201 09:02:25.101440 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7cad829-b344-4e49-b0d4-264750c0ae0f-log-httpd\") pod \"ceilometer-0\" (UID: \"c7cad829-b344-4e49-b0d4-264750c0ae0f\") " pod="openstack/ceilometer-0" Dec 01 09:02:25 crc kubenswrapper[4873]: E1201 09:02:25.146214 4873 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2d169f6e_9eff_40ba_bd22_d38a10f48bd5.slice\": RecentStats: unable to find data in memory cache]" Dec 01 09:02:25 crc kubenswrapper[4873]: I1201 09:02:25.203840 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c7cad829-b344-4e49-b0d4-264750c0ae0f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c7cad829-b344-4e49-b0d4-264750c0ae0f\") " pod="openstack/ceilometer-0" Dec 01 09:02:25 crc kubenswrapper[4873]: I1201 09:02:25.203928 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7cad829-b344-4e49-b0d4-264750c0ae0f-config-data\") pod \"ceilometer-0\" (UID: \"c7cad829-b344-4e49-b0d4-264750c0ae0f\") " pod="openstack/ceilometer-0" Dec 01 09:02:25 crc kubenswrapper[4873]: I1201 09:02:25.203991 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7cad829-b344-4e49-b0d4-264750c0ae0f-run-httpd\") pod \"ceilometer-0\" (UID: \"c7cad829-b344-4e49-b0d4-264750c0ae0f\") " pod="openstack/ceilometer-0" Dec 01 09:02:25 crc kubenswrapper[4873]: I1201 09:02:25.204021 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7cad829-b344-4e49-b0d4-264750c0ae0f-scripts\") pod \"ceilometer-0\" (UID: \"c7cad829-b344-4e49-b0d4-264750c0ae0f\") " pod="openstack/ceilometer-0" Dec 01 09:02:25 crc kubenswrapper[4873]: I1201 09:02:25.204079 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7cad829-b344-4e49-b0d4-264750c0ae0f-log-httpd\") pod \"ceilometer-0\" (UID: \"c7cad829-b344-4e49-b0d4-264750c0ae0f\") " pod="openstack/ceilometer-0" Dec 01 09:02:25 crc kubenswrapper[4873]: I1201 09:02:25.204130 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7srnd\" (UniqueName: \"kubernetes.io/projected/c7cad829-b344-4e49-b0d4-264750c0ae0f-kube-api-access-7srnd\") pod \"ceilometer-0\" (UID: \"c7cad829-b344-4e49-b0d4-264750c0ae0f\") " pod="openstack/ceilometer-0" Dec 01 09:02:25 crc kubenswrapper[4873]: I1201 09:02:25.204148 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7cad829-b344-4e49-b0d4-264750c0ae0f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c7cad829-b344-4e49-b0d4-264750c0ae0f\") " pod="openstack/ceilometer-0" Dec 01 09:02:25 crc kubenswrapper[4873]: I1201 09:02:25.204173 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7cad829-b344-4e49-b0d4-264750c0ae0f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c7cad829-b344-4e49-b0d4-264750c0ae0f\") " pod="openstack/ceilometer-0" Dec 01 09:02:25 crc kubenswrapper[4873]: I1201 09:02:25.205123 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7cad829-b344-4e49-b0d4-264750c0ae0f-run-httpd\") pod \"ceilometer-0\" (UID: \"c7cad829-b344-4e49-b0d4-264750c0ae0f\") " pod="openstack/ceilometer-0" Dec 01 09:02:25 crc kubenswrapper[4873]: I1201 09:02:25.205729 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7cad829-b344-4e49-b0d4-264750c0ae0f-log-httpd\") pod \"ceilometer-0\" (UID: \"c7cad829-b344-4e49-b0d4-264750c0ae0f\") " pod="openstack/ceilometer-0" Dec 01 09:02:25 crc kubenswrapper[4873]: I1201 09:02:25.216391 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7cad829-b344-4e49-b0d4-264750c0ae0f-config-data\") pod \"ceilometer-0\" (UID: \"c7cad829-b344-4e49-b0d4-264750c0ae0f\") " pod="openstack/ceilometer-0" Dec 01 09:02:25 crc kubenswrapper[4873]: I1201 09:02:25.219362 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c7cad829-b344-4e49-b0d4-264750c0ae0f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c7cad829-b344-4e49-b0d4-264750c0ae0f\") " pod="openstack/ceilometer-0" Dec 01 09:02:25 crc kubenswrapper[4873]: I1201 09:02:25.219946 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7cad829-b344-4e49-b0d4-264750c0ae0f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c7cad829-b344-4e49-b0d4-264750c0ae0f\") " pod="openstack/ceilometer-0" Dec 01 09:02:25 crc kubenswrapper[4873]: I1201 09:02:25.222071 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7cad829-b344-4e49-b0d4-264750c0ae0f-scripts\") pod \"ceilometer-0\" (UID: \"c7cad829-b344-4e49-b0d4-264750c0ae0f\") " pod="openstack/ceilometer-0" Dec 01 09:02:25 crc kubenswrapper[4873]: I1201 09:02:25.225348 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7cad829-b344-4e49-b0d4-264750c0ae0f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c7cad829-b344-4e49-b0d4-264750c0ae0f\") " pod="openstack/ceilometer-0" Dec 01 09:02:25 crc kubenswrapper[4873]: I1201 09:02:25.241960 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7srnd\" (UniqueName: \"kubernetes.io/projected/c7cad829-b344-4e49-b0d4-264750c0ae0f-kube-api-access-7srnd\") pod \"ceilometer-0\" (UID: \"c7cad829-b344-4e49-b0d4-264750c0ae0f\") " pod="openstack/ceilometer-0" Dec 01 09:02:25 crc kubenswrapper[4873]: I1201 09:02:25.371239 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 09:02:25 crc kubenswrapper[4873]: I1201 09:02:25.979119 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 09:02:26 crc kubenswrapper[4873]: W1201 09:02:26.051302 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc7cad829_b344_4e49_b0d4_264750c0ae0f.slice/crio-9d6c935168502ee2af9ff782bac01b6f4fa38ed351b738b792e8f36c1190b1b3 WatchSource:0}: Error finding container 9d6c935168502ee2af9ff782bac01b6f4fa38ed351b738b792e8f36c1190b1b3: Status 404 returned error can't find the container with id 9d6c935168502ee2af9ff782bac01b6f4fa38ed351b738b792e8f36c1190b1b3 Dec 01 09:02:26 crc kubenswrapper[4873]: I1201 09:02:26.369511 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 01 09:02:26 crc kubenswrapper[4873]: I1201 09:02:26.369936 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 01 09:02:26 crc kubenswrapper[4873]: I1201 09:02:26.395074 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 01 09:02:26 crc kubenswrapper[4873]: I1201 09:02:26.421642 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 01 09:02:26 crc kubenswrapper[4873]: I1201 09:02:26.430748 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 09:02:26 crc kubenswrapper[4873]: I1201 09:02:26.440567 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d169f6e-9eff-40ba-bd22-d38a10f48bd5" path="/var/lib/kubelet/pods/2d169f6e-9eff-40ba-bd22-d38a10f48bd5/volumes" Dec 01 09:02:26 crc kubenswrapper[4873]: I1201 09:02:26.539178 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80b11b29-b572-4fb2-a32e-7529fbf58637-combined-ca-bundle\") pod \"80b11b29-b572-4fb2-a32e-7529fbf58637\" (UID: \"80b11b29-b572-4fb2-a32e-7529fbf58637\") " Dec 01 09:02:26 crc kubenswrapper[4873]: I1201 09:02:26.539239 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80b11b29-b572-4fb2-a32e-7529fbf58637-logs\") pod \"80b11b29-b572-4fb2-a32e-7529fbf58637\" (UID: \"80b11b29-b572-4fb2-a32e-7529fbf58637\") " Dec 01 09:02:26 crc kubenswrapper[4873]: I1201 09:02:26.539333 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-77xng\" (UniqueName: \"kubernetes.io/projected/80b11b29-b572-4fb2-a32e-7529fbf58637-kube-api-access-77xng\") pod \"80b11b29-b572-4fb2-a32e-7529fbf58637\" (UID: \"80b11b29-b572-4fb2-a32e-7529fbf58637\") " Dec 01 09:02:26 crc kubenswrapper[4873]: I1201 09:02:26.539398 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80b11b29-b572-4fb2-a32e-7529fbf58637-config-data\") pod \"80b11b29-b572-4fb2-a32e-7529fbf58637\" (UID: \"80b11b29-b572-4fb2-a32e-7529fbf58637\") " Dec 01 09:02:26 crc kubenswrapper[4873]: I1201 09:02:26.540491 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80b11b29-b572-4fb2-a32e-7529fbf58637-logs" (OuterVolumeSpecName: "logs") pod "80b11b29-b572-4fb2-a32e-7529fbf58637" (UID: "80b11b29-b572-4fb2-a32e-7529fbf58637"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:02:26 crc kubenswrapper[4873]: I1201 09:02:26.560301 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80b11b29-b572-4fb2-a32e-7529fbf58637-kube-api-access-77xng" (OuterVolumeSpecName: "kube-api-access-77xng") pod "80b11b29-b572-4fb2-a32e-7529fbf58637" (UID: "80b11b29-b572-4fb2-a32e-7529fbf58637"). InnerVolumeSpecName "kube-api-access-77xng". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:02:26 crc kubenswrapper[4873]: I1201 09:02:26.569434 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80b11b29-b572-4fb2-a32e-7529fbf58637-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "80b11b29-b572-4fb2-a32e-7529fbf58637" (UID: "80b11b29-b572-4fb2-a32e-7529fbf58637"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:02:26 crc kubenswrapper[4873]: I1201 09:02:26.576174 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80b11b29-b572-4fb2-a32e-7529fbf58637-config-data" (OuterVolumeSpecName: "config-data") pod "80b11b29-b572-4fb2-a32e-7529fbf58637" (UID: "80b11b29-b572-4fb2-a32e-7529fbf58637"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:02:26 crc kubenswrapper[4873]: I1201 09:02:26.641952 4873 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80b11b29-b572-4fb2-a32e-7529fbf58637-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:26 crc kubenswrapper[4873]: I1201 09:02:26.641999 4873 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80b11b29-b572-4fb2-a32e-7529fbf58637-logs\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:26 crc kubenswrapper[4873]: I1201 09:02:26.642029 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-77xng\" (UniqueName: \"kubernetes.io/projected/80b11b29-b572-4fb2-a32e-7529fbf58637-kube-api-access-77xng\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:26 crc kubenswrapper[4873]: I1201 09:02:26.642047 4873 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80b11b29-b572-4fb2-a32e-7529fbf58637-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:26 crc kubenswrapper[4873]: I1201 09:02:26.861251 4873 generic.go:334] "Generic (PLEG): container finished" podID="80b11b29-b572-4fb2-a32e-7529fbf58637" containerID="c7ccfd4c2e96d8e165124420150043cd20cd95accfc258abd83cd6440a508c18" exitCode=0 Dec 01 09:02:26 crc kubenswrapper[4873]: I1201 09:02:26.861339 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"80b11b29-b572-4fb2-a32e-7529fbf58637","Type":"ContainerDied","Data":"c7ccfd4c2e96d8e165124420150043cd20cd95accfc258abd83cd6440a508c18"} Dec 01 09:02:26 crc kubenswrapper[4873]: I1201 09:02:26.861379 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"80b11b29-b572-4fb2-a32e-7529fbf58637","Type":"ContainerDied","Data":"b1d00334f558ca43a0f491b625035608cff2bdee3d9c57aa25ff901208e54037"} Dec 01 09:02:26 crc kubenswrapper[4873]: I1201 09:02:26.861401 4873 scope.go:117] "RemoveContainer" containerID="c7ccfd4c2e96d8e165124420150043cd20cd95accfc258abd83cd6440a508c18" Dec 01 09:02:26 crc kubenswrapper[4873]: I1201 09:02:26.861526 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 09:02:26 crc kubenswrapper[4873]: I1201 09:02:26.874887 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7cad829-b344-4e49-b0d4-264750c0ae0f","Type":"ContainerStarted","Data":"9d6c935168502ee2af9ff782bac01b6f4fa38ed351b738b792e8f36c1190b1b3"} Dec 01 09:02:26 crc kubenswrapper[4873]: I1201 09:02:26.908730 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 09:02:26 crc kubenswrapper[4873]: I1201 09:02:26.909622 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 01 09:02:26 crc kubenswrapper[4873]: I1201 09:02:26.923556 4873 scope.go:117] "RemoveContainer" containerID="03cb9c7b4a0410bea1842a201e610d35b17a50a977a7b82c45562c3afbf8f908" Dec 01 09:02:26 crc kubenswrapper[4873]: I1201 09:02:26.925685 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 01 09:02:26 crc kubenswrapper[4873]: I1201 09:02:26.938461 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 01 09:02:26 crc kubenswrapper[4873]: E1201 09:02:26.938951 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80b11b29-b572-4fb2-a32e-7529fbf58637" containerName="nova-api-log" Dec 01 09:02:26 crc kubenswrapper[4873]: I1201 09:02:26.938975 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="80b11b29-b572-4fb2-a32e-7529fbf58637" containerName="nova-api-log" Dec 01 09:02:26 crc kubenswrapper[4873]: E1201 09:02:26.938992 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80b11b29-b572-4fb2-a32e-7529fbf58637" containerName="nova-api-api" Dec 01 09:02:26 crc kubenswrapper[4873]: I1201 09:02:26.938999 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="80b11b29-b572-4fb2-a32e-7529fbf58637" containerName="nova-api-api" Dec 01 09:02:26 crc kubenswrapper[4873]: I1201 09:02:26.939311 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="80b11b29-b572-4fb2-a32e-7529fbf58637" containerName="nova-api-log" Dec 01 09:02:26 crc kubenswrapper[4873]: I1201 09:02:26.939342 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="80b11b29-b572-4fb2-a32e-7529fbf58637" containerName="nova-api-api" Dec 01 09:02:26 crc kubenswrapper[4873]: I1201 09:02:26.940777 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 09:02:26 crc kubenswrapper[4873]: I1201 09:02:26.947122 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 01 09:02:26 crc kubenswrapper[4873]: I1201 09:02:26.947269 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 01 09:02:26 crc kubenswrapper[4873]: I1201 09:02:26.947495 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 01 09:02:26 crc kubenswrapper[4873]: I1201 09:02:26.950064 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/61dfd223-55d8-4f16-ad11-280e83f0d358-public-tls-certs\") pod \"nova-api-0\" (UID: \"61dfd223-55d8-4f16-ad11-280e83f0d358\") " pod="openstack/nova-api-0" Dec 01 09:02:26 crc kubenswrapper[4873]: I1201 09:02:26.950170 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61dfd223-55d8-4f16-ad11-280e83f0d358-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"61dfd223-55d8-4f16-ad11-280e83f0d358\") " pod="openstack/nova-api-0" Dec 01 09:02:26 crc kubenswrapper[4873]: I1201 09:02:26.950202 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61dfd223-55d8-4f16-ad11-280e83f0d358-config-data\") pod \"nova-api-0\" (UID: \"61dfd223-55d8-4f16-ad11-280e83f0d358\") " pod="openstack/nova-api-0" Dec 01 09:02:26 crc kubenswrapper[4873]: I1201 09:02:26.950318 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61dfd223-55d8-4f16-ad11-280e83f0d358-logs\") pod \"nova-api-0\" (UID: \"61dfd223-55d8-4f16-ad11-280e83f0d358\") " pod="openstack/nova-api-0" Dec 01 09:02:26 crc kubenswrapper[4873]: I1201 09:02:26.950380 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/61dfd223-55d8-4f16-ad11-280e83f0d358-internal-tls-certs\") pod \"nova-api-0\" (UID: \"61dfd223-55d8-4f16-ad11-280e83f0d358\") " pod="openstack/nova-api-0" Dec 01 09:02:26 crc kubenswrapper[4873]: I1201 09:02:26.950421 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6455p\" (UniqueName: \"kubernetes.io/projected/61dfd223-55d8-4f16-ad11-280e83f0d358-kube-api-access-6455p\") pod \"nova-api-0\" (UID: \"61dfd223-55d8-4f16-ad11-280e83f0d358\") " pod="openstack/nova-api-0" Dec 01 09:02:27 crc kubenswrapper[4873]: I1201 09:02:26.968533 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 09:02:27 crc kubenswrapper[4873]: I1201 09:02:27.003999 4873 scope.go:117] "RemoveContainer" containerID="c7ccfd4c2e96d8e165124420150043cd20cd95accfc258abd83cd6440a508c18" Dec 01 09:02:27 crc kubenswrapper[4873]: E1201 09:02:27.006676 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7ccfd4c2e96d8e165124420150043cd20cd95accfc258abd83cd6440a508c18\": container with ID starting with c7ccfd4c2e96d8e165124420150043cd20cd95accfc258abd83cd6440a508c18 not found: ID does not exist" containerID="c7ccfd4c2e96d8e165124420150043cd20cd95accfc258abd83cd6440a508c18" Dec 01 09:02:27 crc kubenswrapper[4873]: I1201 09:02:27.006730 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7ccfd4c2e96d8e165124420150043cd20cd95accfc258abd83cd6440a508c18"} err="failed to get container status \"c7ccfd4c2e96d8e165124420150043cd20cd95accfc258abd83cd6440a508c18\": rpc error: code = NotFound desc = could not find container \"c7ccfd4c2e96d8e165124420150043cd20cd95accfc258abd83cd6440a508c18\": container with ID starting with c7ccfd4c2e96d8e165124420150043cd20cd95accfc258abd83cd6440a508c18 not found: ID does not exist" Dec 01 09:02:27 crc kubenswrapper[4873]: I1201 09:02:27.006761 4873 scope.go:117] "RemoveContainer" containerID="03cb9c7b4a0410bea1842a201e610d35b17a50a977a7b82c45562c3afbf8f908" Dec 01 09:02:27 crc kubenswrapper[4873]: E1201 09:02:27.014522 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03cb9c7b4a0410bea1842a201e610d35b17a50a977a7b82c45562c3afbf8f908\": container with ID starting with 03cb9c7b4a0410bea1842a201e610d35b17a50a977a7b82c45562c3afbf8f908 not found: ID does not exist" containerID="03cb9c7b4a0410bea1842a201e610d35b17a50a977a7b82c45562c3afbf8f908" Dec 01 09:02:27 crc kubenswrapper[4873]: I1201 09:02:27.014774 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03cb9c7b4a0410bea1842a201e610d35b17a50a977a7b82c45562c3afbf8f908"} err="failed to get container status \"03cb9c7b4a0410bea1842a201e610d35b17a50a977a7b82c45562c3afbf8f908\": rpc error: code = NotFound desc = could not find container \"03cb9c7b4a0410bea1842a201e610d35b17a50a977a7b82c45562c3afbf8f908\": container with ID starting with 03cb9c7b4a0410bea1842a201e610d35b17a50a977a7b82c45562c3afbf8f908 not found: ID does not exist" Dec 01 09:02:27 crc kubenswrapper[4873]: I1201 09:02:27.054585 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/61dfd223-55d8-4f16-ad11-280e83f0d358-public-tls-certs\") pod \"nova-api-0\" (UID: \"61dfd223-55d8-4f16-ad11-280e83f0d358\") " pod="openstack/nova-api-0" Dec 01 09:02:27 crc kubenswrapper[4873]: I1201 09:02:27.054655 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61dfd223-55d8-4f16-ad11-280e83f0d358-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"61dfd223-55d8-4f16-ad11-280e83f0d358\") " pod="openstack/nova-api-0" Dec 01 09:02:27 crc kubenswrapper[4873]: I1201 09:02:27.054696 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61dfd223-55d8-4f16-ad11-280e83f0d358-config-data\") pod \"nova-api-0\" (UID: \"61dfd223-55d8-4f16-ad11-280e83f0d358\") " pod="openstack/nova-api-0" Dec 01 09:02:27 crc kubenswrapper[4873]: I1201 09:02:27.054765 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61dfd223-55d8-4f16-ad11-280e83f0d358-logs\") pod \"nova-api-0\" (UID: \"61dfd223-55d8-4f16-ad11-280e83f0d358\") " pod="openstack/nova-api-0" Dec 01 09:02:27 crc kubenswrapper[4873]: I1201 09:02:27.054797 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/61dfd223-55d8-4f16-ad11-280e83f0d358-internal-tls-certs\") pod \"nova-api-0\" (UID: \"61dfd223-55d8-4f16-ad11-280e83f0d358\") " pod="openstack/nova-api-0" Dec 01 09:02:27 crc kubenswrapper[4873]: I1201 09:02:27.054825 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6455p\" (UniqueName: \"kubernetes.io/projected/61dfd223-55d8-4f16-ad11-280e83f0d358-kube-api-access-6455p\") pod \"nova-api-0\" (UID: \"61dfd223-55d8-4f16-ad11-280e83f0d358\") " pod="openstack/nova-api-0" Dec 01 09:02:27 crc kubenswrapper[4873]: I1201 09:02:27.056203 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61dfd223-55d8-4f16-ad11-280e83f0d358-logs\") pod \"nova-api-0\" (UID: \"61dfd223-55d8-4f16-ad11-280e83f0d358\") " pod="openstack/nova-api-0" Dec 01 09:02:27 crc kubenswrapper[4873]: I1201 09:02:27.066039 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61dfd223-55d8-4f16-ad11-280e83f0d358-config-data\") pod \"nova-api-0\" (UID: \"61dfd223-55d8-4f16-ad11-280e83f0d358\") " pod="openstack/nova-api-0" Dec 01 09:02:27 crc kubenswrapper[4873]: I1201 09:02:27.066813 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61dfd223-55d8-4f16-ad11-280e83f0d358-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"61dfd223-55d8-4f16-ad11-280e83f0d358\") " pod="openstack/nova-api-0" Dec 01 09:02:27 crc kubenswrapper[4873]: I1201 09:02:27.070332 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/61dfd223-55d8-4f16-ad11-280e83f0d358-internal-tls-certs\") pod \"nova-api-0\" (UID: \"61dfd223-55d8-4f16-ad11-280e83f0d358\") " pod="openstack/nova-api-0" Dec 01 09:02:27 crc kubenswrapper[4873]: I1201 09:02:27.070935 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/61dfd223-55d8-4f16-ad11-280e83f0d358-public-tls-certs\") pod \"nova-api-0\" (UID: \"61dfd223-55d8-4f16-ad11-280e83f0d358\") " pod="openstack/nova-api-0" Dec 01 09:02:27 crc kubenswrapper[4873]: I1201 09:02:27.096227 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6455p\" (UniqueName: \"kubernetes.io/projected/61dfd223-55d8-4f16-ad11-280e83f0d358-kube-api-access-6455p\") pod \"nova-api-0\" (UID: \"61dfd223-55d8-4f16-ad11-280e83f0d358\") " pod="openstack/nova-api-0" Dec 01 09:02:27 crc kubenswrapper[4873]: I1201 09:02:27.142663 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-h4n9r"] Dec 01 09:02:27 crc kubenswrapper[4873]: I1201 09:02:27.144722 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-h4n9r" Dec 01 09:02:27 crc kubenswrapper[4873]: I1201 09:02:27.149667 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 01 09:02:27 crc kubenswrapper[4873]: I1201 09:02:27.153250 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 01 09:02:27 crc kubenswrapper[4873]: I1201 09:02:27.204132 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-h4n9r"] Dec 01 09:02:27 crc kubenswrapper[4873]: I1201 09:02:27.262286 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a911632-e929-4993-9f93-f88d6e6c8993-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-h4n9r\" (UID: \"0a911632-e929-4993-9f93-f88d6e6c8993\") " pod="openstack/nova-cell1-cell-mapping-h4n9r" Dec 01 09:02:27 crc kubenswrapper[4873]: I1201 09:02:27.263257 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a911632-e929-4993-9f93-f88d6e6c8993-config-data\") pod \"nova-cell1-cell-mapping-h4n9r\" (UID: \"0a911632-e929-4993-9f93-f88d6e6c8993\") " pod="openstack/nova-cell1-cell-mapping-h4n9r" Dec 01 09:02:27 crc kubenswrapper[4873]: I1201 09:02:27.263676 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6j6z\" (UniqueName: \"kubernetes.io/projected/0a911632-e929-4993-9f93-f88d6e6c8993-kube-api-access-d6j6z\") pod \"nova-cell1-cell-mapping-h4n9r\" (UID: \"0a911632-e929-4993-9f93-f88d6e6c8993\") " pod="openstack/nova-cell1-cell-mapping-h4n9r" Dec 01 09:02:27 crc kubenswrapper[4873]: I1201 09:02:27.264198 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a911632-e929-4993-9f93-f88d6e6c8993-scripts\") pod \"nova-cell1-cell-mapping-h4n9r\" (UID: \"0a911632-e929-4993-9f93-f88d6e6c8993\") " pod="openstack/nova-cell1-cell-mapping-h4n9r" Dec 01 09:02:27 crc kubenswrapper[4873]: I1201 09:02:27.311151 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 09:02:27 crc kubenswrapper[4873]: I1201 09:02:27.366814 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a911632-e929-4993-9f93-f88d6e6c8993-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-h4n9r\" (UID: \"0a911632-e929-4993-9f93-f88d6e6c8993\") " pod="openstack/nova-cell1-cell-mapping-h4n9r" Dec 01 09:02:27 crc kubenswrapper[4873]: I1201 09:02:27.366884 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a911632-e929-4993-9f93-f88d6e6c8993-config-data\") pod \"nova-cell1-cell-mapping-h4n9r\" (UID: \"0a911632-e929-4993-9f93-f88d6e6c8993\") " pod="openstack/nova-cell1-cell-mapping-h4n9r" Dec 01 09:02:27 crc kubenswrapper[4873]: I1201 09:02:27.367243 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6j6z\" (UniqueName: \"kubernetes.io/projected/0a911632-e929-4993-9f93-f88d6e6c8993-kube-api-access-d6j6z\") pod \"nova-cell1-cell-mapping-h4n9r\" (UID: \"0a911632-e929-4993-9f93-f88d6e6c8993\") " pod="openstack/nova-cell1-cell-mapping-h4n9r" Dec 01 09:02:27 crc kubenswrapper[4873]: I1201 09:02:27.367328 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a911632-e929-4993-9f93-f88d6e6c8993-scripts\") pod \"nova-cell1-cell-mapping-h4n9r\" (UID: \"0a911632-e929-4993-9f93-f88d6e6c8993\") " pod="openstack/nova-cell1-cell-mapping-h4n9r" Dec 01 09:02:27 crc kubenswrapper[4873]: I1201 09:02:27.375729 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a911632-e929-4993-9f93-f88d6e6c8993-config-data\") pod \"nova-cell1-cell-mapping-h4n9r\" (UID: \"0a911632-e929-4993-9f93-f88d6e6c8993\") " pod="openstack/nova-cell1-cell-mapping-h4n9r" Dec 01 09:02:27 crc kubenswrapper[4873]: I1201 09:02:27.375751 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a911632-e929-4993-9f93-f88d6e6c8993-scripts\") pod \"nova-cell1-cell-mapping-h4n9r\" (UID: \"0a911632-e929-4993-9f93-f88d6e6c8993\") " pod="openstack/nova-cell1-cell-mapping-h4n9r" Dec 01 09:02:27 crc kubenswrapper[4873]: I1201 09:02:27.378490 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a911632-e929-4993-9f93-f88d6e6c8993-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-h4n9r\" (UID: \"0a911632-e929-4993-9f93-f88d6e6c8993\") " pod="openstack/nova-cell1-cell-mapping-h4n9r" Dec 01 09:02:27 crc kubenswrapper[4873]: I1201 09:02:27.383369 4873 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="50f5253a-ab03-426c-aaca-a8db33425fef" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.179:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 01 09:02:27 crc kubenswrapper[4873]: I1201 09:02:27.383747 4873 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="50f5253a-ab03-426c-aaca-a8db33425fef" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.179:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 01 09:02:27 crc kubenswrapper[4873]: I1201 09:02:27.391538 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6j6z\" (UniqueName: \"kubernetes.io/projected/0a911632-e929-4993-9f93-f88d6e6c8993-kube-api-access-d6j6z\") pod \"nova-cell1-cell-mapping-h4n9r\" (UID: \"0a911632-e929-4993-9f93-f88d6e6c8993\") " pod="openstack/nova-cell1-cell-mapping-h4n9r" Dec 01 09:02:27 crc kubenswrapper[4873]: I1201 09:02:27.507590 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-h4n9r" Dec 01 09:02:27 crc kubenswrapper[4873]: I1201 09:02:27.877649 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 09:02:27 crc kubenswrapper[4873]: I1201 09:02:27.888628 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"61dfd223-55d8-4f16-ad11-280e83f0d358","Type":"ContainerStarted","Data":"dc76037044375c163f725daef1261843ff937a0732080bcb35df9215ab1cd200"} Dec 01 09:02:27 crc kubenswrapper[4873]: I1201 09:02:27.892071 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7cad829-b344-4e49-b0d4-264750c0ae0f","Type":"ContainerStarted","Data":"f168c0536d36e29586d04755ace417b659bf6cc437bea36b669ff4b731d89e24"} Dec 01 09:02:28 crc kubenswrapper[4873]: I1201 09:02:28.052302 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-h4n9r"] Dec 01 09:02:28 crc kubenswrapper[4873]: I1201 09:02:28.446153 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80b11b29-b572-4fb2-a32e-7529fbf58637" path="/var/lib/kubelet/pods/80b11b29-b572-4fb2-a32e-7529fbf58637/volumes" Dec 01 09:02:28 crc kubenswrapper[4873]: I1201 09:02:28.905736 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-h4n9r" event={"ID":"0a911632-e929-4993-9f93-f88d6e6c8993","Type":"ContainerStarted","Data":"410b00412bb25258ee0029e32b927f1d745f3d5ab7fb21f04a4bd75930426a37"} Dec 01 09:02:28 crc kubenswrapper[4873]: I1201 09:02:28.905794 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-h4n9r" event={"ID":"0a911632-e929-4993-9f93-f88d6e6c8993","Type":"ContainerStarted","Data":"20353445f09ca98cd23897c7f6c6973e7735b8bfcf493382865ecbe0c09715dd"} Dec 01 09:02:28 crc kubenswrapper[4873]: I1201 09:02:28.908443 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"61dfd223-55d8-4f16-ad11-280e83f0d358","Type":"ContainerStarted","Data":"46169d05da15793c53520a0e12c47736830f3622655e63e58f4d6ce158442b0d"} Dec 01 09:02:28 crc kubenswrapper[4873]: I1201 09:02:28.908494 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"61dfd223-55d8-4f16-ad11-280e83f0d358","Type":"ContainerStarted","Data":"bac152f1bcd70b6fd6effa739e7c71ee2382f4aade85521951fd0211e07f854e"} Dec 01 09:02:28 crc kubenswrapper[4873]: I1201 09:02:28.910661 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7cad829-b344-4e49-b0d4-264750c0ae0f","Type":"ContainerStarted","Data":"515feff768ee635a4bc61549e094ad2a0d6362338a61af5ceeedf238ceb452c4"} Dec 01 09:02:28 crc kubenswrapper[4873]: I1201 09:02:28.935585 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-h4n9r" podStartSLOduration=1.93555215 podStartE2EDuration="1.93555215s" podCreationTimestamp="2025-12-01 09:02:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:02:28.924560871 +0000 UTC m=+1324.826669410" watchObservedRunningTime="2025-12-01 09:02:28.93555215 +0000 UTC m=+1324.837660719" Dec 01 09:02:28 crc kubenswrapper[4873]: I1201 09:02:28.959107 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.959069355 podStartE2EDuration="2.959069355s" podCreationTimestamp="2025-12-01 09:02:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:02:28.954441312 +0000 UTC m=+1324.856549851" watchObservedRunningTime="2025-12-01 09:02:28.959069355 +0000 UTC m=+1324.861177904" Dec 01 09:02:29 crc kubenswrapper[4873]: I1201 09:02:29.923464 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7cad829-b344-4e49-b0d4-264750c0ae0f","Type":"ContainerStarted","Data":"c91515ebd87bb1e28b9586fe19335671f8d956b8ecdad460712529b7d0846130"} Dec 01 09:02:30 crc kubenswrapper[4873]: I1201 09:02:30.361258 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b856c5697-rw4vp" Dec 01 09:02:30 crc kubenswrapper[4873]: I1201 09:02:30.481661 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-566b5b7845-zxctn"] Dec 01 09:02:30 crc kubenswrapper[4873]: I1201 09:02:30.482071 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-566b5b7845-zxctn" podUID="cd1d4f43-ecf0-4716-8491-9f322329f2c9" containerName="dnsmasq-dns" containerID="cri-o://9d1768aa02375c0c2a6f07207bc68d7fbc7b61220569fa7396770fed5cb31f2c" gracePeriod=10 Dec 01 09:02:30 crc kubenswrapper[4873]: I1201 09:02:30.954737 4873 generic.go:334] "Generic (PLEG): container finished" podID="cd1d4f43-ecf0-4716-8491-9f322329f2c9" containerID="9d1768aa02375c0c2a6f07207bc68d7fbc7b61220569fa7396770fed5cb31f2c" exitCode=0 Dec 01 09:02:30 crc kubenswrapper[4873]: I1201 09:02:30.954832 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-566b5b7845-zxctn" event={"ID":"cd1d4f43-ecf0-4716-8491-9f322329f2c9","Type":"ContainerDied","Data":"9d1768aa02375c0c2a6f07207bc68d7fbc7b61220569fa7396770fed5cb31f2c"} Dec 01 09:02:30 crc kubenswrapper[4873]: I1201 09:02:30.955597 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-566b5b7845-zxctn" event={"ID":"cd1d4f43-ecf0-4716-8491-9f322329f2c9","Type":"ContainerDied","Data":"d9b8c7daa1e2153464cd87447548d826c16d1677384960ba9d8673696e2bf1fd"} Dec 01 09:02:30 crc kubenswrapper[4873]: I1201 09:02:30.955617 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d9b8c7daa1e2153464cd87447548d826c16d1677384960ba9d8673696e2bf1fd" Dec 01 09:02:31 crc kubenswrapper[4873]: I1201 09:02:31.063540 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:02:31 crc kubenswrapper[4873]: I1201 09:02:31.063636 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:02:31 crc kubenswrapper[4873]: I1201 09:02:31.064155 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-566b5b7845-zxctn" Dec 01 09:02:31 crc kubenswrapper[4873]: I1201 09:02:31.102414 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cd1d4f43-ecf0-4716-8491-9f322329f2c9-ovsdbserver-sb\") pod \"cd1d4f43-ecf0-4716-8491-9f322329f2c9\" (UID: \"cd1d4f43-ecf0-4716-8491-9f322329f2c9\") " Dec 01 09:02:31 crc kubenswrapper[4873]: I1201 09:02:31.102513 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd1d4f43-ecf0-4716-8491-9f322329f2c9-config\") pod \"cd1d4f43-ecf0-4716-8491-9f322329f2c9\" (UID: \"cd1d4f43-ecf0-4716-8491-9f322329f2c9\") " Dec 01 09:02:31 crc kubenswrapper[4873]: I1201 09:02:31.102707 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vk596\" (UniqueName: \"kubernetes.io/projected/cd1d4f43-ecf0-4716-8491-9f322329f2c9-kube-api-access-vk596\") pod \"cd1d4f43-ecf0-4716-8491-9f322329f2c9\" (UID: \"cd1d4f43-ecf0-4716-8491-9f322329f2c9\") " Dec 01 09:02:31 crc kubenswrapper[4873]: I1201 09:02:31.102990 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cd1d4f43-ecf0-4716-8491-9f322329f2c9-ovsdbserver-nb\") pod \"cd1d4f43-ecf0-4716-8491-9f322329f2c9\" (UID: \"cd1d4f43-ecf0-4716-8491-9f322329f2c9\") " Dec 01 09:02:31 crc kubenswrapper[4873]: I1201 09:02:31.103137 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cd1d4f43-ecf0-4716-8491-9f322329f2c9-dns-svc\") pod \"cd1d4f43-ecf0-4716-8491-9f322329f2c9\" (UID: \"cd1d4f43-ecf0-4716-8491-9f322329f2c9\") " Dec 01 09:02:31 crc kubenswrapper[4873]: I1201 09:02:31.122082 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd1d4f43-ecf0-4716-8491-9f322329f2c9-kube-api-access-vk596" (OuterVolumeSpecName: "kube-api-access-vk596") pod "cd1d4f43-ecf0-4716-8491-9f322329f2c9" (UID: "cd1d4f43-ecf0-4716-8491-9f322329f2c9"). InnerVolumeSpecName "kube-api-access-vk596". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:02:31 crc kubenswrapper[4873]: I1201 09:02:31.199862 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd1d4f43-ecf0-4716-8491-9f322329f2c9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cd1d4f43-ecf0-4716-8491-9f322329f2c9" (UID: "cd1d4f43-ecf0-4716-8491-9f322329f2c9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:02:31 crc kubenswrapper[4873]: I1201 09:02:31.200385 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd1d4f43-ecf0-4716-8491-9f322329f2c9-config" (OuterVolumeSpecName: "config") pod "cd1d4f43-ecf0-4716-8491-9f322329f2c9" (UID: "cd1d4f43-ecf0-4716-8491-9f322329f2c9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:02:31 crc kubenswrapper[4873]: I1201 09:02:31.201252 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd1d4f43-ecf0-4716-8491-9f322329f2c9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "cd1d4f43-ecf0-4716-8491-9f322329f2c9" (UID: "cd1d4f43-ecf0-4716-8491-9f322329f2c9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:02:31 crc kubenswrapper[4873]: I1201 09:02:31.205060 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd1d4f43-ecf0-4716-8491-9f322329f2c9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "cd1d4f43-ecf0-4716-8491-9f322329f2c9" (UID: "cd1d4f43-ecf0-4716-8491-9f322329f2c9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:02:31 crc kubenswrapper[4873]: I1201 09:02:31.210377 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vk596\" (UniqueName: \"kubernetes.io/projected/cd1d4f43-ecf0-4716-8491-9f322329f2c9-kube-api-access-vk596\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:31 crc kubenswrapper[4873]: I1201 09:02:31.210514 4873 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cd1d4f43-ecf0-4716-8491-9f322329f2c9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:31 crc kubenswrapper[4873]: I1201 09:02:31.210677 4873 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cd1d4f43-ecf0-4716-8491-9f322329f2c9-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:31 crc kubenswrapper[4873]: I1201 09:02:31.210743 4873 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cd1d4f43-ecf0-4716-8491-9f322329f2c9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:31 crc kubenswrapper[4873]: I1201 09:02:31.210800 4873 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd1d4f43-ecf0-4716-8491-9f322329f2c9-config\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:31 crc kubenswrapper[4873]: I1201 09:02:31.970289 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-566b5b7845-zxctn" Dec 01 09:02:31 crc kubenswrapper[4873]: I1201 09:02:31.970459 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7cad829-b344-4e49-b0d4-264750c0ae0f","Type":"ContainerStarted","Data":"2bd108f1fad71742ecc62b281d05995aefb751321590b194fa66df222271b3eb"} Dec 01 09:02:31 crc kubenswrapper[4873]: I1201 09:02:31.974062 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 09:02:32 crc kubenswrapper[4873]: I1201 09:02:32.006208 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.170775681 podStartE2EDuration="8.006176296s" podCreationTimestamp="2025-12-01 09:02:24 +0000 UTC" firstStartedPulling="2025-12-01 09:02:26.057185106 +0000 UTC m=+1321.959293635" lastFinishedPulling="2025-12-01 09:02:30.892585701 +0000 UTC m=+1326.794694250" observedRunningTime="2025-12-01 09:02:31.989121088 +0000 UTC m=+1327.891229637" watchObservedRunningTime="2025-12-01 09:02:32.006176296 +0000 UTC m=+1327.908284835" Dec 01 09:02:32 crc kubenswrapper[4873]: I1201 09:02:32.025247 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-566b5b7845-zxctn"] Dec 01 09:02:32 crc kubenswrapper[4873]: I1201 09:02:32.033519 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-566b5b7845-zxctn"] Dec 01 09:02:32 crc kubenswrapper[4873]: I1201 09:02:32.449239 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd1d4f43-ecf0-4716-8491-9f322329f2c9" path="/var/lib/kubelet/pods/cd1d4f43-ecf0-4716-8491-9f322329f2c9/volumes" Dec 01 09:02:35 crc kubenswrapper[4873]: I1201 09:02:35.012897 4873 generic.go:334] "Generic (PLEG): container finished" podID="0a911632-e929-4993-9f93-f88d6e6c8993" containerID="410b00412bb25258ee0029e32b927f1d745f3d5ab7fb21f04a4bd75930426a37" exitCode=0 Dec 01 09:02:35 crc kubenswrapper[4873]: I1201 09:02:35.013003 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-h4n9r" event={"ID":"0a911632-e929-4993-9f93-f88d6e6c8993","Type":"ContainerDied","Data":"410b00412bb25258ee0029e32b927f1d745f3d5ab7fb21f04a4bd75930426a37"} Dec 01 09:02:36 crc kubenswrapper[4873]: I1201 09:02:36.376741 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 01 09:02:36 crc kubenswrapper[4873]: I1201 09:02:36.386170 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 01 09:02:36 crc kubenswrapper[4873]: I1201 09:02:36.387563 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 01 09:02:36 crc kubenswrapper[4873]: I1201 09:02:36.408065 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-h4n9r" Dec 01 09:02:36 crc kubenswrapper[4873]: I1201 09:02:36.536936 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a911632-e929-4993-9f93-f88d6e6c8993-config-data\") pod \"0a911632-e929-4993-9f93-f88d6e6c8993\" (UID: \"0a911632-e929-4993-9f93-f88d6e6c8993\") " Dec 01 09:02:36 crc kubenswrapper[4873]: I1201 09:02:36.537267 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a911632-e929-4993-9f93-f88d6e6c8993-scripts\") pod \"0a911632-e929-4993-9f93-f88d6e6c8993\" (UID: \"0a911632-e929-4993-9f93-f88d6e6c8993\") " Dec 01 09:02:36 crc kubenswrapper[4873]: I1201 09:02:36.537297 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a911632-e929-4993-9f93-f88d6e6c8993-combined-ca-bundle\") pod \"0a911632-e929-4993-9f93-f88d6e6c8993\" (UID: \"0a911632-e929-4993-9f93-f88d6e6c8993\") " Dec 01 09:02:36 crc kubenswrapper[4873]: I1201 09:02:36.537380 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6j6z\" (UniqueName: \"kubernetes.io/projected/0a911632-e929-4993-9f93-f88d6e6c8993-kube-api-access-d6j6z\") pod \"0a911632-e929-4993-9f93-f88d6e6c8993\" (UID: \"0a911632-e929-4993-9f93-f88d6e6c8993\") " Dec 01 09:02:36 crc kubenswrapper[4873]: I1201 09:02:36.543955 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a911632-e929-4993-9f93-f88d6e6c8993-scripts" (OuterVolumeSpecName: "scripts") pod "0a911632-e929-4993-9f93-f88d6e6c8993" (UID: "0a911632-e929-4993-9f93-f88d6e6c8993"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:02:36 crc kubenswrapper[4873]: I1201 09:02:36.544033 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a911632-e929-4993-9f93-f88d6e6c8993-kube-api-access-d6j6z" (OuterVolumeSpecName: "kube-api-access-d6j6z") pod "0a911632-e929-4993-9f93-f88d6e6c8993" (UID: "0a911632-e929-4993-9f93-f88d6e6c8993"). InnerVolumeSpecName "kube-api-access-d6j6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:02:36 crc kubenswrapper[4873]: I1201 09:02:36.566999 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a911632-e929-4993-9f93-f88d6e6c8993-config-data" (OuterVolumeSpecName: "config-data") pod "0a911632-e929-4993-9f93-f88d6e6c8993" (UID: "0a911632-e929-4993-9f93-f88d6e6c8993"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:02:36 crc kubenswrapper[4873]: I1201 09:02:36.569485 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a911632-e929-4993-9f93-f88d6e6c8993-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0a911632-e929-4993-9f93-f88d6e6c8993" (UID: "0a911632-e929-4993-9f93-f88d6e6c8993"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:02:36 crc kubenswrapper[4873]: I1201 09:02:36.640245 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6j6z\" (UniqueName: \"kubernetes.io/projected/0a911632-e929-4993-9f93-f88d6e6c8993-kube-api-access-d6j6z\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:36 crc kubenswrapper[4873]: I1201 09:02:36.640300 4873 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a911632-e929-4993-9f93-f88d6e6c8993-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:36 crc kubenswrapper[4873]: I1201 09:02:36.640310 4873 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a911632-e929-4993-9f93-f88d6e6c8993-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:36 crc kubenswrapper[4873]: I1201 09:02:36.640321 4873 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a911632-e929-4993-9f93-f88d6e6c8993-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:37 crc kubenswrapper[4873]: I1201 09:02:37.035564 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-h4n9r" Dec 01 09:02:37 crc kubenswrapper[4873]: I1201 09:02:37.035566 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-h4n9r" event={"ID":"0a911632-e929-4993-9f93-f88d6e6c8993","Type":"ContainerDied","Data":"20353445f09ca98cd23897c7f6c6973e7735b8bfcf493382865ecbe0c09715dd"} Dec 01 09:02:37 crc kubenswrapper[4873]: I1201 09:02:37.035617 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="20353445f09ca98cd23897c7f6c6973e7735b8bfcf493382865ecbe0c09715dd" Dec 01 09:02:37 crc kubenswrapper[4873]: I1201 09:02:37.046440 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 01 09:02:37 crc kubenswrapper[4873]: I1201 09:02:37.231357 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 09:02:37 crc kubenswrapper[4873]: I1201 09:02:37.231677 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="61dfd223-55d8-4f16-ad11-280e83f0d358" containerName="nova-api-log" containerID="cri-o://bac152f1bcd70b6fd6effa739e7c71ee2382f4aade85521951fd0211e07f854e" gracePeriod=30 Dec 01 09:02:37 crc kubenswrapper[4873]: I1201 09:02:37.231777 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="61dfd223-55d8-4f16-ad11-280e83f0d358" containerName="nova-api-api" containerID="cri-o://46169d05da15793c53520a0e12c47736830f3622655e63e58f4d6ce158442b0d" gracePeriod=30 Dec 01 09:02:37 crc kubenswrapper[4873]: I1201 09:02:37.295470 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 09:02:37 crc kubenswrapper[4873]: I1201 09:02:37.296234 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="8dce570f-2670-49a0-a42c-8e12d118f279" containerName="nova-scheduler-scheduler" containerID="cri-o://fe777aad3bb04934c08d3f70a2b914656cc6fe44825115f3b18a8b4e1faa1d81" gracePeriod=30 Dec 01 09:02:37 crc kubenswrapper[4873]: I1201 09:02:37.309544 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 09:02:37 crc kubenswrapper[4873]: I1201 09:02:37.842313 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 09:02:37 crc kubenswrapper[4873]: E1201 09:02:37.922409 4873 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="fe777aad3bb04934c08d3f70a2b914656cc6fe44825115f3b18a8b4e1faa1d81" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 01 09:02:37 crc kubenswrapper[4873]: E1201 09:02:37.924241 4873 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="fe777aad3bb04934c08d3f70a2b914656cc6fe44825115f3b18a8b4e1faa1d81" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 01 09:02:37 crc kubenswrapper[4873]: E1201 09:02:37.925304 4873 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="fe777aad3bb04934c08d3f70a2b914656cc6fe44825115f3b18a8b4e1faa1d81" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 01 09:02:37 crc kubenswrapper[4873]: E1201 09:02:37.925355 4873 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="8dce570f-2670-49a0-a42c-8e12d118f279" containerName="nova-scheduler-scheduler" Dec 01 09:02:37 crc kubenswrapper[4873]: I1201 09:02:37.989738 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61dfd223-55d8-4f16-ad11-280e83f0d358-config-data\") pod \"61dfd223-55d8-4f16-ad11-280e83f0d358\" (UID: \"61dfd223-55d8-4f16-ad11-280e83f0d358\") " Dec 01 09:02:37 crc kubenswrapper[4873]: I1201 09:02:37.989816 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61dfd223-55d8-4f16-ad11-280e83f0d358-combined-ca-bundle\") pod \"61dfd223-55d8-4f16-ad11-280e83f0d358\" (UID: \"61dfd223-55d8-4f16-ad11-280e83f0d358\") " Dec 01 09:02:37 crc kubenswrapper[4873]: I1201 09:02:37.989945 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/61dfd223-55d8-4f16-ad11-280e83f0d358-internal-tls-certs\") pod \"61dfd223-55d8-4f16-ad11-280e83f0d358\" (UID: \"61dfd223-55d8-4f16-ad11-280e83f0d358\") " Dec 01 09:02:37 crc kubenswrapper[4873]: I1201 09:02:37.990062 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61dfd223-55d8-4f16-ad11-280e83f0d358-logs\") pod \"61dfd223-55d8-4f16-ad11-280e83f0d358\" (UID: \"61dfd223-55d8-4f16-ad11-280e83f0d358\") " Dec 01 09:02:37 crc kubenswrapper[4873]: I1201 09:02:37.990148 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6455p\" (UniqueName: \"kubernetes.io/projected/61dfd223-55d8-4f16-ad11-280e83f0d358-kube-api-access-6455p\") pod \"61dfd223-55d8-4f16-ad11-280e83f0d358\" (UID: \"61dfd223-55d8-4f16-ad11-280e83f0d358\") " Dec 01 09:02:37 crc kubenswrapper[4873]: I1201 09:02:37.990260 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/61dfd223-55d8-4f16-ad11-280e83f0d358-public-tls-certs\") pod \"61dfd223-55d8-4f16-ad11-280e83f0d358\" (UID: \"61dfd223-55d8-4f16-ad11-280e83f0d358\") " Dec 01 09:02:37 crc kubenswrapper[4873]: I1201 09:02:37.990947 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61dfd223-55d8-4f16-ad11-280e83f0d358-logs" (OuterVolumeSpecName: "logs") pod "61dfd223-55d8-4f16-ad11-280e83f0d358" (UID: "61dfd223-55d8-4f16-ad11-280e83f0d358"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:02:37 crc kubenswrapper[4873]: I1201 09:02:37.992003 4873 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61dfd223-55d8-4f16-ad11-280e83f0d358-logs\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.005392 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61dfd223-55d8-4f16-ad11-280e83f0d358-kube-api-access-6455p" (OuterVolumeSpecName: "kube-api-access-6455p") pod "61dfd223-55d8-4f16-ad11-280e83f0d358" (UID: "61dfd223-55d8-4f16-ad11-280e83f0d358"). InnerVolumeSpecName "kube-api-access-6455p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.022969 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61dfd223-55d8-4f16-ad11-280e83f0d358-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "61dfd223-55d8-4f16-ad11-280e83f0d358" (UID: "61dfd223-55d8-4f16-ad11-280e83f0d358"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.037523 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61dfd223-55d8-4f16-ad11-280e83f0d358-config-data" (OuterVolumeSpecName: "config-data") pod "61dfd223-55d8-4f16-ad11-280e83f0d358" (UID: "61dfd223-55d8-4f16-ad11-280e83f0d358"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.048875 4873 generic.go:334] "Generic (PLEG): container finished" podID="61dfd223-55d8-4f16-ad11-280e83f0d358" containerID="46169d05da15793c53520a0e12c47736830f3622655e63e58f4d6ce158442b0d" exitCode=0 Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.048927 4873 generic.go:334] "Generic (PLEG): container finished" podID="61dfd223-55d8-4f16-ad11-280e83f0d358" containerID="bac152f1bcd70b6fd6effa739e7c71ee2382f4aade85521951fd0211e07f854e" exitCode=143 Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.049495 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.049576 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"61dfd223-55d8-4f16-ad11-280e83f0d358","Type":"ContainerDied","Data":"46169d05da15793c53520a0e12c47736830f3622655e63e58f4d6ce158442b0d"} Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.049668 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"61dfd223-55d8-4f16-ad11-280e83f0d358","Type":"ContainerDied","Data":"bac152f1bcd70b6fd6effa739e7c71ee2382f4aade85521951fd0211e07f854e"} Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.049687 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"61dfd223-55d8-4f16-ad11-280e83f0d358","Type":"ContainerDied","Data":"dc76037044375c163f725daef1261843ff937a0732080bcb35df9215ab1cd200"} Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.049710 4873 scope.go:117] "RemoveContainer" containerID="46169d05da15793c53520a0e12c47736830f3622655e63e58f4d6ce158442b0d" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.058751 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61dfd223-55d8-4f16-ad11-280e83f0d358-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "61dfd223-55d8-4f16-ad11-280e83f0d358" (UID: "61dfd223-55d8-4f16-ad11-280e83f0d358"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.075193 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61dfd223-55d8-4f16-ad11-280e83f0d358-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "61dfd223-55d8-4f16-ad11-280e83f0d358" (UID: "61dfd223-55d8-4f16-ad11-280e83f0d358"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.097334 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6455p\" (UniqueName: \"kubernetes.io/projected/61dfd223-55d8-4f16-ad11-280e83f0d358-kube-api-access-6455p\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.097374 4873 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/61dfd223-55d8-4f16-ad11-280e83f0d358-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.097385 4873 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61dfd223-55d8-4f16-ad11-280e83f0d358-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.097396 4873 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61dfd223-55d8-4f16-ad11-280e83f0d358-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.097406 4873 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/61dfd223-55d8-4f16-ad11-280e83f0d358-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.110606 4873 scope.go:117] "RemoveContainer" containerID="bac152f1bcd70b6fd6effa739e7c71ee2382f4aade85521951fd0211e07f854e" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.137932 4873 scope.go:117] "RemoveContainer" containerID="46169d05da15793c53520a0e12c47736830f3622655e63e58f4d6ce158442b0d" Dec 01 09:02:38 crc kubenswrapper[4873]: E1201 09:02:38.138579 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46169d05da15793c53520a0e12c47736830f3622655e63e58f4d6ce158442b0d\": container with ID starting with 46169d05da15793c53520a0e12c47736830f3622655e63e58f4d6ce158442b0d not found: ID does not exist" containerID="46169d05da15793c53520a0e12c47736830f3622655e63e58f4d6ce158442b0d" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.138637 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46169d05da15793c53520a0e12c47736830f3622655e63e58f4d6ce158442b0d"} err="failed to get container status \"46169d05da15793c53520a0e12c47736830f3622655e63e58f4d6ce158442b0d\": rpc error: code = NotFound desc = could not find container \"46169d05da15793c53520a0e12c47736830f3622655e63e58f4d6ce158442b0d\": container with ID starting with 46169d05da15793c53520a0e12c47736830f3622655e63e58f4d6ce158442b0d not found: ID does not exist" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.138669 4873 scope.go:117] "RemoveContainer" containerID="bac152f1bcd70b6fd6effa739e7c71ee2382f4aade85521951fd0211e07f854e" Dec 01 09:02:38 crc kubenswrapper[4873]: E1201 09:02:38.139055 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bac152f1bcd70b6fd6effa739e7c71ee2382f4aade85521951fd0211e07f854e\": container with ID starting with bac152f1bcd70b6fd6effa739e7c71ee2382f4aade85521951fd0211e07f854e not found: ID does not exist" containerID="bac152f1bcd70b6fd6effa739e7c71ee2382f4aade85521951fd0211e07f854e" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.139112 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bac152f1bcd70b6fd6effa739e7c71ee2382f4aade85521951fd0211e07f854e"} err="failed to get container status \"bac152f1bcd70b6fd6effa739e7c71ee2382f4aade85521951fd0211e07f854e\": rpc error: code = NotFound desc = could not find container \"bac152f1bcd70b6fd6effa739e7c71ee2382f4aade85521951fd0211e07f854e\": container with ID starting with bac152f1bcd70b6fd6effa739e7c71ee2382f4aade85521951fd0211e07f854e not found: ID does not exist" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.139151 4873 scope.go:117] "RemoveContainer" containerID="46169d05da15793c53520a0e12c47736830f3622655e63e58f4d6ce158442b0d" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.139484 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46169d05da15793c53520a0e12c47736830f3622655e63e58f4d6ce158442b0d"} err="failed to get container status \"46169d05da15793c53520a0e12c47736830f3622655e63e58f4d6ce158442b0d\": rpc error: code = NotFound desc = could not find container \"46169d05da15793c53520a0e12c47736830f3622655e63e58f4d6ce158442b0d\": container with ID starting with 46169d05da15793c53520a0e12c47736830f3622655e63e58f4d6ce158442b0d not found: ID does not exist" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.139508 4873 scope.go:117] "RemoveContainer" containerID="bac152f1bcd70b6fd6effa739e7c71ee2382f4aade85521951fd0211e07f854e" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.139996 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bac152f1bcd70b6fd6effa739e7c71ee2382f4aade85521951fd0211e07f854e"} err="failed to get container status \"bac152f1bcd70b6fd6effa739e7c71ee2382f4aade85521951fd0211e07f854e\": rpc error: code = NotFound desc = could not find container \"bac152f1bcd70b6fd6effa739e7c71ee2382f4aade85521951fd0211e07f854e\": container with ID starting with bac152f1bcd70b6fd6effa739e7c71ee2382f4aade85521951fd0211e07f854e not found: ID does not exist" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.429271 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.464915 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.475563 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 01 09:02:38 crc kubenswrapper[4873]: E1201 09:02:38.476077 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61dfd223-55d8-4f16-ad11-280e83f0d358" containerName="nova-api-log" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.476100 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="61dfd223-55d8-4f16-ad11-280e83f0d358" containerName="nova-api-log" Dec 01 09:02:38 crc kubenswrapper[4873]: E1201 09:02:38.476131 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd1d4f43-ecf0-4716-8491-9f322329f2c9" containerName="init" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.476139 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd1d4f43-ecf0-4716-8491-9f322329f2c9" containerName="init" Dec 01 09:02:38 crc kubenswrapper[4873]: E1201 09:02:38.476154 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61dfd223-55d8-4f16-ad11-280e83f0d358" containerName="nova-api-api" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.476162 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="61dfd223-55d8-4f16-ad11-280e83f0d358" containerName="nova-api-api" Dec 01 09:02:38 crc kubenswrapper[4873]: E1201 09:02:38.476177 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a911632-e929-4993-9f93-f88d6e6c8993" containerName="nova-manage" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.476187 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a911632-e929-4993-9f93-f88d6e6c8993" containerName="nova-manage" Dec 01 09:02:38 crc kubenswrapper[4873]: E1201 09:02:38.476202 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd1d4f43-ecf0-4716-8491-9f322329f2c9" containerName="dnsmasq-dns" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.476222 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd1d4f43-ecf0-4716-8491-9f322329f2c9" containerName="dnsmasq-dns" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.476400 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a911632-e929-4993-9f93-f88d6e6c8993" containerName="nova-manage" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.476419 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="61dfd223-55d8-4f16-ad11-280e83f0d358" containerName="nova-api-api" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.476436 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd1d4f43-ecf0-4716-8491-9f322329f2c9" containerName="dnsmasq-dns" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.476447 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="61dfd223-55d8-4f16-ad11-280e83f0d358" containerName="nova-api-log" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.477744 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.485289 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.488514 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.488825 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.488846 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.609339 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/11dc97a8-cb16-4992-abc5-4fe5de83bac0-public-tls-certs\") pod \"nova-api-0\" (UID: \"11dc97a8-cb16-4992-abc5-4fe5de83bac0\") " pod="openstack/nova-api-0" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.615592 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/11dc97a8-cb16-4992-abc5-4fe5de83bac0-internal-tls-certs\") pod \"nova-api-0\" (UID: \"11dc97a8-cb16-4992-abc5-4fe5de83bac0\") " pod="openstack/nova-api-0" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.615661 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/11dc97a8-cb16-4992-abc5-4fe5de83bac0-logs\") pod \"nova-api-0\" (UID: \"11dc97a8-cb16-4992-abc5-4fe5de83bac0\") " pod="openstack/nova-api-0" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.615909 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11dc97a8-cb16-4992-abc5-4fe5de83bac0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"11dc97a8-cb16-4992-abc5-4fe5de83bac0\") " pod="openstack/nova-api-0" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.616179 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cb5qj\" (UniqueName: \"kubernetes.io/projected/11dc97a8-cb16-4992-abc5-4fe5de83bac0-kube-api-access-cb5qj\") pod \"nova-api-0\" (UID: \"11dc97a8-cb16-4992-abc5-4fe5de83bac0\") " pod="openstack/nova-api-0" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.616543 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11dc97a8-cb16-4992-abc5-4fe5de83bac0-config-data\") pod \"nova-api-0\" (UID: \"11dc97a8-cb16-4992-abc5-4fe5de83bac0\") " pod="openstack/nova-api-0" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.718924 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/11dc97a8-cb16-4992-abc5-4fe5de83bac0-public-tls-certs\") pod \"nova-api-0\" (UID: \"11dc97a8-cb16-4992-abc5-4fe5de83bac0\") " pod="openstack/nova-api-0" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.719066 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/11dc97a8-cb16-4992-abc5-4fe5de83bac0-internal-tls-certs\") pod \"nova-api-0\" (UID: \"11dc97a8-cb16-4992-abc5-4fe5de83bac0\") " pod="openstack/nova-api-0" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.719099 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/11dc97a8-cb16-4992-abc5-4fe5de83bac0-logs\") pod \"nova-api-0\" (UID: \"11dc97a8-cb16-4992-abc5-4fe5de83bac0\") " pod="openstack/nova-api-0" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.719136 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11dc97a8-cb16-4992-abc5-4fe5de83bac0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"11dc97a8-cb16-4992-abc5-4fe5de83bac0\") " pod="openstack/nova-api-0" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.719197 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cb5qj\" (UniqueName: \"kubernetes.io/projected/11dc97a8-cb16-4992-abc5-4fe5de83bac0-kube-api-access-cb5qj\") pod \"nova-api-0\" (UID: \"11dc97a8-cb16-4992-abc5-4fe5de83bac0\") " pod="openstack/nova-api-0" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.719299 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11dc97a8-cb16-4992-abc5-4fe5de83bac0-config-data\") pod \"nova-api-0\" (UID: \"11dc97a8-cb16-4992-abc5-4fe5de83bac0\") " pod="openstack/nova-api-0" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.719762 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/11dc97a8-cb16-4992-abc5-4fe5de83bac0-logs\") pod \"nova-api-0\" (UID: \"11dc97a8-cb16-4992-abc5-4fe5de83bac0\") " pod="openstack/nova-api-0" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.724833 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/11dc97a8-cb16-4992-abc5-4fe5de83bac0-public-tls-certs\") pod \"nova-api-0\" (UID: \"11dc97a8-cb16-4992-abc5-4fe5de83bac0\") " pod="openstack/nova-api-0" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.724838 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/11dc97a8-cb16-4992-abc5-4fe5de83bac0-internal-tls-certs\") pod \"nova-api-0\" (UID: \"11dc97a8-cb16-4992-abc5-4fe5de83bac0\") " pod="openstack/nova-api-0" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.725196 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11dc97a8-cb16-4992-abc5-4fe5de83bac0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"11dc97a8-cb16-4992-abc5-4fe5de83bac0\") " pod="openstack/nova-api-0" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.725618 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11dc97a8-cb16-4992-abc5-4fe5de83bac0-config-data\") pod \"nova-api-0\" (UID: \"11dc97a8-cb16-4992-abc5-4fe5de83bac0\") " pod="openstack/nova-api-0" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.746104 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cb5qj\" (UniqueName: \"kubernetes.io/projected/11dc97a8-cb16-4992-abc5-4fe5de83bac0-kube-api-access-cb5qj\") pod \"nova-api-0\" (UID: \"11dc97a8-cb16-4992-abc5-4fe5de83bac0\") " pod="openstack/nova-api-0" Dec 01 09:02:38 crc kubenswrapper[4873]: I1201 09:02:38.801447 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 09:02:39 crc kubenswrapper[4873]: I1201 09:02:39.065219 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="50f5253a-ab03-426c-aaca-a8db33425fef" containerName="nova-metadata-metadata" containerID="cri-o://50276f38fa8798d2887c903bee97a511cc88f6de965d5963d74b1f8e57edb89e" gracePeriod=30 Dec 01 09:02:39 crc kubenswrapper[4873]: I1201 09:02:39.065363 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="50f5253a-ab03-426c-aaca-a8db33425fef" containerName="nova-metadata-log" containerID="cri-o://b962fe9f343506d6931f2e98b12c658595cee840435c60e66d8a85e3351d0dd4" gracePeriod=30 Dec 01 09:02:39 crc kubenswrapper[4873]: I1201 09:02:39.278814 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 09:02:39 crc kubenswrapper[4873]: W1201 09:02:39.286909 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod11dc97a8_cb16_4992_abc5_4fe5de83bac0.slice/crio-736215f38fb1d92c57b352a5afdaf4c3d194e6eb6f9dc499e794ae11734e7144 WatchSource:0}: Error finding container 736215f38fb1d92c57b352a5afdaf4c3d194e6eb6f9dc499e794ae11734e7144: Status 404 returned error can't find the container with id 736215f38fb1d92c57b352a5afdaf4c3d194e6eb6f9dc499e794ae11734e7144 Dec 01 09:02:40 crc kubenswrapper[4873]: I1201 09:02:40.080646 4873 generic.go:334] "Generic (PLEG): container finished" podID="50f5253a-ab03-426c-aaca-a8db33425fef" containerID="b962fe9f343506d6931f2e98b12c658595cee840435c60e66d8a85e3351d0dd4" exitCode=143 Dec 01 09:02:40 crc kubenswrapper[4873]: I1201 09:02:40.080870 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"50f5253a-ab03-426c-aaca-a8db33425fef","Type":"ContainerDied","Data":"b962fe9f343506d6931f2e98b12c658595cee840435c60e66d8a85e3351d0dd4"} Dec 01 09:02:40 crc kubenswrapper[4873]: I1201 09:02:40.084389 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"11dc97a8-cb16-4992-abc5-4fe5de83bac0","Type":"ContainerStarted","Data":"25494f7b5662e5de0537d252567634278e9d3a719c63ae0a73d866b16fb5e944"} Dec 01 09:02:40 crc kubenswrapper[4873]: I1201 09:02:40.084423 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"11dc97a8-cb16-4992-abc5-4fe5de83bac0","Type":"ContainerStarted","Data":"c21e78f3dc013d992342d9af2547747442a601ad8daeda46f6b5716118aa9ace"} Dec 01 09:02:40 crc kubenswrapper[4873]: I1201 09:02:40.084435 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"11dc97a8-cb16-4992-abc5-4fe5de83bac0","Type":"ContainerStarted","Data":"736215f38fb1d92c57b352a5afdaf4c3d194e6eb6f9dc499e794ae11734e7144"} Dec 01 09:02:40 crc kubenswrapper[4873]: I1201 09:02:40.110880 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.110860656 podStartE2EDuration="2.110860656s" podCreationTimestamp="2025-12-01 09:02:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:02:40.107272218 +0000 UTC m=+1336.009380777" watchObservedRunningTime="2025-12-01 09:02:40.110860656 +0000 UTC m=+1336.012969195" Dec 01 09:02:40 crc kubenswrapper[4873]: I1201 09:02:40.443500 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61dfd223-55d8-4f16-ad11-280e83f0d358" path="/var/lib/kubelet/pods/61dfd223-55d8-4f16-ad11-280e83f0d358/volumes" Dec 01 09:02:42 crc kubenswrapper[4873]: I1201 09:02:42.131075 4873 generic.go:334] "Generic (PLEG): container finished" podID="8dce570f-2670-49a0-a42c-8e12d118f279" containerID="fe777aad3bb04934c08d3f70a2b914656cc6fe44825115f3b18a8b4e1faa1d81" exitCode=0 Dec 01 09:02:42 crc kubenswrapper[4873]: I1201 09:02:42.131531 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8dce570f-2670-49a0-a42c-8e12d118f279","Type":"ContainerDied","Data":"fe777aad3bb04934c08d3f70a2b914656cc6fe44825115f3b18a8b4e1faa1d81"} Dec 01 09:02:42 crc kubenswrapper[4873]: I1201 09:02:42.247223 4873 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="50f5253a-ab03-426c-aaca-a8db33425fef" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.179:8775/\": read tcp 10.217.0.2:44592->10.217.0.179:8775: read: connection reset by peer" Dec 01 09:02:42 crc kubenswrapper[4873]: I1201 09:02:42.247223 4873 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="50f5253a-ab03-426c-aaca-a8db33425fef" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.179:8775/\": read tcp 10.217.0.2:44594->10.217.0.179:8775: read: connection reset by peer" Dec 01 09:02:42 crc kubenswrapper[4873]: I1201 09:02:42.381895 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 09:02:42 crc kubenswrapper[4873]: I1201 09:02:42.414109 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dce570f-2670-49a0-a42c-8e12d118f279-combined-ca-bundle\") pod \"8dce570f-2670-49a0-a42c-8e12d118f279\" (UID: \"8dce570f-2670-49a0-a42c-8e12d118f279\") " Dec 01 09:02:42 crc kubenswrapper[4873]: I1201 09:02:42.414245 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w5gd2\" (UniqueName: \"kubernetes.io/projected/8dce570f-2670-49a0-a42c-8e12d118f279-kube-api-access-w5gd2\") pod \"8dce570f-2670-49a0-a42c-8e12d118f279\" (UID: \"8dce570f-2670-49a0-a42c-8e12d118f279\") " Dec 01 09:02:42 crc kubenswrapper[4873]: I1201 09:02:42.435357 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dce570f-2670-49a0-a42c-8e12d118f279-kube-api-access-w5gd2" (OuterVolumeSpecName: "kube-api-access-w5gd2") pod "8dce570f-2670-49a0-a42c-8e12d118f279" (UID: "8dce570f-2670-49a0-a42c-8e12d118f279"). InnerVolumeSpecName "kube-api-access-w5gd2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:02:42 crc kubenswrapper[4873]: I1201 09:02:42.456945 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dce570f-2670-49a0-a42c-8e12d118f279-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8dce570f-2670-49a0-a42c-8e12d118f279" (UID: "8dce570f-2670-49a0-a42c-8e12d118f279"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:02:42 crc kubenswrapper[4873]: I1201 09:02:42.524171 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8dce570f-2670-49a0-a42c-8e12d118f279-config-data\") pod \"8dce570f-2670-49a0-a42c-8e12d118f279\" (UID: \"8dce570f-2670-49a0-a42c-8e12d118f279\") " Dec 01 09:02:42 crc kubenswrapper[4873]: I1201 09:02:42.525624 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w5gd2\" (UniqueName: \"kubernetes.io/projected/8dce570f-2670-49a0-a42c-8e12d118f279-kube-api-access-w5gd2\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:42 crc kubenswrapper[4873]: I1201 09:02:42.525654 4873 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dce570f-2670-49a0-a42c-8e12d118f279-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:42 crc kubenswrapper[4873]: I1201 09:02:42.570473 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dce570f-2670-49a0-a42c-8e12d118f279-config-data" (OuterVolumeSpecName: "config-data") pod "8dce570f-2670-49a0-a42c-8e12d118f279" (UID: "8dce570f-2670-49a0-a42c-8e12d118f279"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:02:42 crc kubenswrapper[4873]: I1201 09:02:42.628866 4873 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8dce570f-2670-49a0-a42c-8e12d118f279-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:42 crc kubenswrapper[4873]: I1201 09:02:42.641691 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 09:02:42 crc kubenswrapper[4873]: I1201 09:02:42.832430 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/50f5253a-ab03-426c-aaca-a8db33425fef-nova-metadata-tls-certs\") pod \"50f5253a-ab03-426c-aaca-a8db33425fef\" (UID: \"50f5253a-ab03-426c-aaca-a8db33425fef\") " Dec 01 09:02:42 crc kubenswrapper[4873]: I1201 09:02:42.832480 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7jdj\" (UniqueName: \"kubernetes.io/projected/50f5253a-ab03-426c-aaca-a8db33425fef-kube-api-access-r7jdj\") pod \"50f5253a-ab03-426c-aaca-a8db33425fef\" (UID: \"50f5253a-ab03-426c-aaca-a8db33425fef\") " Dec 01 09:02:42 crc kubenswrapper[4873]: I1201 09:02:42.832601 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50f5253a-ab03-426c-aaca-a8db33425fef-logs\") pod \"50f5253a-ab03-426c-aaca-a8db33425fef\" (UID: \"50f5253a-ab03-426c-aaca-a8db33425fef\") " Dec 01 09:02:42 crc kubenswrapper[4873]: I1201 09:02:42.832810 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50f5253a-ab03-426c-aaca-a8db33425fef-combined-ca-bundle\") pod \"50f5253a-ab03-426c-aaca-a8db33425fef\" (UID: \"50f5253a-ab03-426c-aaca-a8db33425fef\") " Dec 01 09:02:42 crc kubenswrapper[4873]: I1201 09:02:42.832885 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50f5253a-ab03-426c-aaca-a8db33425fef-config-data\") pod \"50f5253a-ab03-426c-aaca-a8db33425fef\" (UID: \"50f5253a-ab03-426c-aaca-a8db33425fef\") " Dec 01 09:02:42 crc kubenswrapper[4873]: I1201 09:02:42.833406 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50f5253a-ab03-426c-aaca-a8db33425fef-logs" (OuterVolumeSpecName: "logs") pod "50f5253a-ab03-426c-aaca-a8db33425fef" (UID: "50f5253a-ab03-426c-aaca-a8db33425fef"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:02:42 crc kubenswrapper[4873]: I1201 09:02:42.835769 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50f5253a-ab03-426c-aaca-a8db33425fef-kube-api-access-r7jdj" (OuterVolumeSpecName: "kube-api-access-r7jdj") pod "50f5253a-ab03-426c-aaca-a8db33425fef" (UID: "50f5253a-ab03-426c-aaca-a8db33425fef"). InnerVolumeSpecName "kube-api-access-r7jdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:02:42 crc kubenswrapper[4873]: I1201 09:02:42.864443 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50f5253a-ab03-426c-aaca-a8db33425fef-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "50f5253a-ab03-426c-aaca-a8db33425fef" (UID: "50f5253a-ab03-426c-aaca-a8db33425fef"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:02:42 crc kubenswrapper[4873]: I1201 09:02:42.867221 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50f5253a-ab03-426c-aaca-a8db33425fef-config-data" (OuterVolumeSpecName: "config-data") pod "50f5253a-ab03-426c-aaca-a8db33425fef" (UID: "50f5253a-ab03-426c-aaca-a8db33425fef"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:02:42 crc kubenswrapper[4873]: I1201 09:02:42.907834 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50f5253a-ab03-426c-aaca-a8db33425fef-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "50f5253a-ab03-426c-aaca-a8db33425fef" (UID: "50f5253a-ab03-426c-aaca-a8db33425fef"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:02:42 crc kubenswrapper[4873]: I1201 09:02:42.935718 4873 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50f5253a-ab03-426c-aaca-a8db33425fef-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:42 crc kubenswrapper[4873]: I1201 09:02:42.935795 4873 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50f5253a-ab03-426c-aaca-a8db33425fef-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:42 crc kubenswrapper[4873]: I1201 09:02:42.935805 4873 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/50f5253a-ab03-426c-aaca-a8db33425fef-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:42 crc kubenswrapper[4873]: I1201 09:02:42.935906 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7jdj\" (UniqueName: \"kubernetes.io/projected/50f5253a-ab03-426c-aaca-a8db33425fef-kube-api-access-r7jdj\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:42 crc kubenswrapper[4873]: I1201 09:02:42.935919 4873 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50f5253a-ab03-426c-aaca-a8db33425fef-logs\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.145979 4873 generic.go:334] "Generic (PLEG): container finished" podID="50f5253a-ab03-426c-aaca-a8db33425fef" containerID="50276f38fa8798d2887c903bee97a511cc88f6de965d5963d74b1f8e57edb89e" exitCode=0 Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.146056 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"50f5253a-ab03-426c-aaca-a8db33425fef","Type":"ContainerDied","Data":"50276f38fa8798d2887c903bee97a511cc88f6de965d5963d74b1f8e57edb89e"} Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.146114 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.146130 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"50f5253a-ab03-426c-aaca-a8db33425fef","Type":"ContainerDied","Data":"14ef3625991cc6aa6b15c1ec6d4d095b2b4d41d5d13793f0331f1e64076e5239"} Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.146158 4873 scope.go:117] "RemoveContainer" containerID="50276f38fa8798d2887c903bee97a511cc88f6de965d5963d74b1f8e57edb89e" Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.153193 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8dce570f-2670-49a0-a42c-8e12d118f279","Type":"ContainerDied","Data":"30d6f7e463dede8b1db5f34b9b6dabbc767e2d4843b3cc782c205f86c3e2db63"} Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.153294 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.188433 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.202245 4873 scope.go:117] "RemoveContainer" containerID="b962fe9f343506d6931f2e98b12c658595cee840435c60e66d8a85e3351d0dd4" Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.204922 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.260590 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 01 09:02:43 crc kubenswrapper[4873]: E1201 09:02:43.261541 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50f5253a-ab03-426c-aaca-a8db33425fef" containerName="nova-metadata-log" Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.261556 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="50f5253a-ab03-426c-aaca-a8db33425fef" containerName="nova-metadata-log" Dec 01 09:02:43 crc kubenswrapper[4873]: E1201 09:02:43.261577 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50f5253a-ab03-426c-aaca-a8db33425fef" containerName="nova-metadata-metadata" Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.261584 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="50f5253a-ab03-426c-aaca-a8db33425fef" containerName="nova-metadata-metadata" Dec 01 09:02:43 crc kubenswrapper[4873]: E1201 09:02:43.261599 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dce570f-2670-49a0-a42c-8e12d118f279" containerName="nova-scheduler-scheduler" Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.261607 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dce570f-2670-49a0-a42c-8e12d118f279" containerName="nova-scheduler-scheduler" Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.261846 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="50f5253a-ab03-426c-aaca-a8db33425fef" containerName="nova-metadata-log" Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.261861 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dce570f-2670-49a0-a42c-8e12d118f279" containerName="nova-scheduler-scheduler" Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.261872 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="50f5253a-ab03-426c-aaca-a8db33425fef" containerName="nova-metadata-metadata" Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.263068 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.263087 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.263217 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.272355 4873 scope.go:117] "RemoveContainer" containerID="50276f38fa8798d2887c903bee97a511cc88f6de965d5963d74b1f8e57edb89e" Dec 01 09:02:43 crc kubenswrapper[4873]: E1201 09:02:43.273644 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50276f38fa8798d2887c903bee97a511cc88f6de965d5963d74b1f8e57edb89e\": container with ID starting with 50276f38fa8798d2887c903bee97a511cc88f6de965d5963d74b1f8e57edb89e not found: ID does not exist" containerID="50276f38fa8798d2887c903bee97a511cc88f6de965d5963d74b1f8e57edb89e" Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.273686 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50276f38fa8798d2887c903bee97a511cc88f6de965d5963d74b1f8e57edb89e"} err="failed to get container status \"50276f38fa8798d2887c903bee97a511cc88f6de965d5963d74b1f8e57edb89e\": rpc error: code = NotFound desc = could not find container \"50276f38fa8798d2887c903bee97a511cc88f6de965d5963d74b1f8e57edb89e\": container with ID starting with 50276f38fa8798d2887c903bee97a511cc88f6de965d5963d74b1f8e57edb89e not found: ID does not exist" Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.273712 4873 scope.go:117] "RemoveContainer" containerID="b962fe9f343506d6931f2e98b12c658595cee840435c60e66d8a85e3351d0dd4" Dec 01 09:02:43 crc kubenswrapper[4873]: E1201 09:02:43.275172 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b962fe9f343506d6931f2e98b12c658595cee840435c60e66d8a85e3351d0dd4\": container with ID starting with b962fe9f343506d6931f2e98b12c658595cee840435c60e66d8a85e3351d0dd4 not found: ID does not exist" containerID="b962fe9f343506d6931f2e98b12c658595cee840435c60e66d8a85e3351d0dd4" Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.275251 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b962fe9f343506d6931f2e98b12c658595cee840435c60e66d8a85e3351d0dd4"} err="failed to get container status \"b962fe9f343506d6931f2e98b12c658595cee840435c60e66d8a85e3351d0dd4\": rpc error: code = NotFound desc = could not find container \"b962fe9f343506d6931f2e98b12c658595cee840435c60e66d8a85e3351d0dd4\": container with ID starting with b962fe9f343506d6931f2e98b12c658595cee840435c60e66d8a85e3351d0dd4 not found: ID does not exist" Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.275294 4873 scope.go:117] "RemoveContainer" containerID="fe777aad3bb04934c08d3f70a2b914656cc6fe44825115f3b18a8b4e1faa1d81" Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.275507 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.276222 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nhbl\" (UniqueName: \"kubernetes.io/projected/e54ecbbf-ac5c-4924-a396-549da1855e65-kube-api-access-7nhbl\") pod \"nova-metadata-0\" (UID: \"e54ecbbf-ac5c-4924-a396-549da1855e65\") " pod="openstack/nova-metadata-0" Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.276272 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e54ecbbf-ac5c-4924-a396-549da1855e65-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e54ecbbf-ac5c-4924-a396-549da1855e65\") " pod="openstack/nova-metadata-0" Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.276339 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e54ecbbf-ac5c-4924-a396-549da1855e65-logs\") pod \"nova-metadata-0\" (UID: \"e54ecbbf-ac5c-4924-a396-549da1855e65\") " pod="openstack/nova-metadata-0" Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.276375 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e54ecbbf-ac5c-4924-a396-549da1855e65-config-data\") pod \"nova-metadata-0\" (UID: \"e54ecbbf-ac5c-4924-a396-549da1855e65\") " pod="openstack/nova-metadata-0" Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.276410 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e54ecbbf-ac5c-4924-a396-549da1855e65-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e54ecbbf-ac5c-4924-a396-549da1855e65\") " pod="openstack/nova-metadata-0" Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.282809 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.283081 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.287473 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.290123 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.294522 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.316984 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.378438 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e54ecbbf-ac5c-4924-a396-549da1855e65-logs\") pod \"nova-metadata-0\" (UID: \"e54ecbbf-ac5c-4924-a396-549da1855e65\") " pod="openstack/nova-metadata-0" Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.378520 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e54ecbbf-ac5c-4924-a396-549da1855e65-config-data\") pod \"nova-metadata-0\" (UID: \"e54ecbbf-ac5c-4924-a396-549da1855e65\") " pod="openstack/nova-metadata-0" Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.378556 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbxn8\" (UniqueName: \"kubernetes.io/projected/634610e2-dd0b-46f5-a934-e7f44249c731-kube-api-access-zbxn8\") pod \"nova-scheduler-0\" (UID: \"634610e2-dd0b-46f5-a934-e7f44249c731\") " pod="openstack/nova-scheduler-0" Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.378582 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e54ecbbf-ac5c-4924-a396-549da1855e65-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e54ecbbf-ac5c-4924-a396-549da1855e65\") " pod="openstack/nova-metadata-0" Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.378613 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/634610e2-dd0b-46f5-a934-e7f44249c731-config-data\") pod \"nova-scheduler-0\" (UID: \"634610e2-dd0b-46f5-a934-e7f44249c731\") " pod="openstack/nova-scheduler-0" Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.378648 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/634610e2-dd0b-46f5-a934-e7f44249c731-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"634610e2-dd0b-46f5-a934-e7f44249c731\") " pod="openstack/nova-scheduler-0" Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.378700 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nhbl\" (UniqueName: \"kubernetes.io/projected/e54ecbbf-ac5c-4924-a396-549da1855e65-kube-api-access-7nhbl\") pod \"nova-metadata-0\" (UID: \"e54ecbbf-ac5c-4924-a396-549da1855e65\") " pod="openstack/nova-metadata-0" Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.378745 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e54ecbbf-ac5c-4924-a396-549da1855e65-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e54ecbbf-ac5c-4924-a396-549da1855e65\") " pod="openstack/nova-metadata-0" Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.379337 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e54ecbbf-ac5c-4924-a396-549da1855e65-logs\") pod \"nova-metadata-0\" (UID: \"e54ecbbf-ac5c-4924-a396-549da1855e65\") " pod="openstack/nova-metadata-0" Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.384388 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e54ecbbf-ac5c-4924-a396-549da1855e65-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e54ecbbf-ac5c-4924-a396-549da1855e65\") " pod="openstack/nova-metadata-0" Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.385805 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e54ecbbf-ac5c-4924-a396-549da1855e65-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e54ecbbf-ac5c-4924-a396-549da1855e65\") " pod="openstack/nova-metadata-0" Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.387568 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e54ecbbf-ac5c-4924-a396-549da1855e65-config-data\") pod \"nova-metadata-0\" (UID: \"e54ecbbf-ac5c-4924-a396-549da1855e65\") " pod="openstack/nova-metadata-0" Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.399393 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nhbl\" (UniqueName: \"kubernetes.io/projected/e54ecbbf-ac5c-4924-a396-549da1855e65-kube-api-access-7nhbl\") pod \"nova-metadata-0\" (UID: \"e54ecbbf-ac5c-4924-a396-549da1855e65\") " pod="openstack/nova-metadata-0" Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.480900 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbxn8\" (UniqueName: \"kubernetes.io/projected/634610e2-dd0b-46f5-a934-e7f44249c731-kube-api-access-zbxn8\") pod \"nova-scheduler-0\" (UID: \"634610e2-dd0b-46f5-a934-e7f44249c731\") " pod="openstack/nova-scheduler-0" Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.481065 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/634610e2-dd0b-46f5-a934-e7f44249c731-config-data\") pod \"nova-scheduler-0\" (UID: \"634610e2-dd0b-46f5-a934-e7f44249c731\") " pod="openstack/nova-scheduler-0" Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.481154 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/634610e2-dd0b-46f5-a934-e7f44249c731-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"634610e2-dd0b-46f5-a934-e7f44249c731\") " pod="openstack/nova-scheduler-0" Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.487303 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/634610e2-dd0b-46f5-a934-e7f44249c731-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"634610e2-dd0b-46f5-a934-e7f44249c731\") " pod="openstack/nova-scheduler-0" Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.487305 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/634610e2-dd0b-46f5-a934-e7f44249c731-config-data\") pod \"nova-scheduler-0\" (UID: \"634610e2-dd0b-46f5-a934-e7f44249c731\") " pod="openstack/nova-scheduler-0" Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.500593 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbxn8\" (UniqueName: \"kubernetes.io/projected/634610e2-dd0b-46f5-a934-e7f44249c731-kube-api-access-zbxn8\") pod \"nova-scheduler-0\" (UID: \"634610e2-dd0b-46f5-a934-e7f44249c731\") " pod="openstack/nova-scheduler-0" Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.600411 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 09:02:43 crc kubenswrapper[4873]: I1201 09:02:43.627772 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 09:02:44 crc kubenswrapper[4873]: I1201 09:02:44.128336 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 09:02:44 crc kubenswrapper[4873]: I1201 09:02:44.173470 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e54ecbbf-ac5c-4924-a396-549da1855e65","Type":"ContainerStarted","Data":"35ba66d22db46767fa83d8949c7a58200718bd3fabc22160b52c9a9c6ba087d7"} Dec 01 09:02:44 crc kubenswrapper[4873]: I1201 09:02:44.213865 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 09:02:44 crc kubenswrapper[4873]: W1201 09:02:44.216307 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod634610e2_dd0b_46f5_a934_e7f44249c731.slice/crio-2cc101d6aa6856300c77a12ad5d8e19a9428dbb43b231f3e2c43e6cb3361546f WatchSource:0}: Error finding container 2cc101d6aa6856300c77a12ad5d8e19a9428dbb43b231f3e2c43e6cb3361546f: Status 404 returned error can't find the container with id 2cc101d6aa6856300c77a12ad5d8e19a9428dbb43b231f3e2c43e6cb3361546f Dec 01 09:02:44 crc kubenswrapper[4873]: I1201 09:02:44.445173 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50f5253a-ab03-426c-aaca-a8db33425fef" path="/var/lib/kubelet/pods/50f5253a-ab03-426c-aaca-a8db33425fef/volumes" Dec 01 09:02:44 crc kubenswrapper[4873]: I1201 09:02:44.445833 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8dce570f-2670-49a0-a42c-8e12d118f279" path="/var/lib/kubelet/pods/8dce570f-2670-49a0-a42c-8e12d118f279/volumes" Dec 01 09:02:45 crc kubenswrapper[4873]: I1201 09:02:45.201286 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"634610e2-dd0b-46f5-a934-e7f44249c731","Type":"ContainerStarted","Data":"b3030fe5ed72933130dd4275a24c67a196678f14dfa81a2e74b8ed99c6ebfb35"} Dec 01 09:02:45 crc kubenswrapper[4873]: I1201 09:02:45.201371 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"634610e2-dd0b-46f5-a934-e7f44249c731","Type":"ContainerStarted","Data":"2cc101d6aa6856300c77a12ad5d8e19a9428dbb43b231f3e2c43e6cb3361546f"} Dec 01 09:02:45 crc kubenswrapper[4873]: I1201 09:02:45.203730 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e54ecbbf-ac5c-4924-a396-549da1855e65","Type":"ContainerStarted","Data":"cf149b0a776a07aaf3609920d2baefc7b9a80b423d737155f67dd2fb1886aebf"} Dec 01 09:02:45 crc kubenswrapper[4873]: I1201 09:02:45.203793 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e54ecbbf-ac5c-4924-a396-549da1855e65","Type":"ContainerStarted","Data":"fe13ca0ef20a66cbcd86a3b282f48a7d48b2fa06f28e380aa609d781760a4178"} Dec 01 09:02:45 crc kubenswrapper[4873]: I1201 09:02:45.229167 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.22914234 podStartE2EDuration="2.22914234s" podCreationTimestamp="2025-12-01 09:02:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:02:45.220863848 +0000 UTC m=+1341.122972397" watchObservedRunningTime="2025-12-01 09:02:45.22914234 +0000 UTC m=+1341.131250879" Dec 01 09:02:45 crc kubenswrapper[4873]: I1201 09:02:45.250371 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.250343619 podStartE2EDuration="2.250343619s" podCreationTimestamp="2025-12-01 09:02:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:02:45.245406048 +0000 UTC m=+1341.147514587" watchObservedRunningTime="2025-12-01 09:02:45.250343619 +0000 UTC m=+1341.152452158" Dec 01 09:02:48 crc kubenswrapper[4873]: I1201 09:02:48.600946 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 01 09:02:48 crc kubenswrapper[4873]: I1201 09:02:48.601528 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 01 09:02:48 crc kubenswrapper[4873]: I1201 09:02:48.627940 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 01 09:02:48 crc kubenswrapper[4873]: I1201 09:02:48.802205 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 09:02:48 crc kubenswrapper[4873]: I1201 09:02:48.802285 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 09:02:49 crc kubenswrapper[4873]: I1201 09:02:49.820553 4873 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="11dc97a8-cb16-4992-abc5-4fe5de83bac0" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.185:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 01 09:02:49 crc kubenswrapper[4873]: I1201 09:02:49.820613 4873 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="11dc97a8-cb16-4992-abc5-4fe5de83bac0" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.185:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 01 09:02:53 crc kubenswrapper[4873]: I1201 09:02:53.602002 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 01 09:02:53 crc kubenswrapper[4873]: I1201 09:02:53.602156 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 01 09:02:53 crc kubenswrapper[4873]: I1201 09:02:53.631200 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 01 09:02:53 crc kubenswrapper[4873]: I1201 09:02:53.677063 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 01 09:02:54 crc kubenswrapper[4873]: I1201 09:02:54.341242 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 01 09:02:54 crc kubenswrapper[4873]: I1201 09:02:54.618308 4873 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="e54ecbbf-ac5c-4924-a396-549da1855e65" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.186:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 01 09:02:54 crc kubenswrapper[4873]: I1201 09:02:54.618330 4873 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="e54ecbbf-ac5c-4924-a396-549da1855e65" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.186:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 01 09:02:55 crc kubenswrapper[4873]: I1201 09:02:55.566905 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 01 09:02:58 crc kubenswrapper[4873]: I1201 09:02:58.819437 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 01 09:02:58 crc kubenswrapper[4873]: I1201 09:02:58.822671 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 01 09:02:58 crc kubenswrapper[4873]: I1201 09:02:58.827102 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 01 09:02:58 crc kubenswrapper[4873]: I1201 09:02:58.831300 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 01 09:02:59 crc kubenswrapper[4873]: I1201 09:02:59.620980 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 01 09:02:59 crc kubenswrapper[4873]: I1201 09:02:59.634744 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 01 09:03:01 crc kubenswrapper[4873]: I1201 09:03:01.059855 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:03:01 crc kubenswrapper[4873]: I1201 09:03:01.059946 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:03:01 crc kubenswrapper[4873]: I1201 09:03:01.060039 4873 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" Dec 01 09:03:01 crc kubenswrapper[4873]: I1201 09:03:01.061338 4873 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"90066b4e910f0ee21a175275171e35716a5459a2c723385e352a837980db5011"} pod="openshift-machine-config-operator/machine-config-daemon-scwpp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 09:03:01 crc kubenswrapper[4873]: I1201 09:03:01.061956 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" containerID="cri-o://90066b4e910f0ee21a175275171e35716a5459a2c723385e352a837980db5011" gracePeriod=600 Dec 01 09:03:01 crc kubenswrapper[4873]: I1201 09:03:01.642166 4873 generic.go:334] "Generic (PLEG): container finished" podID="fef7b114-0e07-402d-a37b-315c36011f4b" containerID="90066b4e910f0ee21a175275171e35716a5459a2c723385e352a837980db5011" exitCode=0 Dec 01 09:03:01 crc kubenswrapper[4873]: I1201 09:03:01.642229 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" event={"ID":"fef7b114-0e07-402d-a37b-315c36011f4b","Type":"ContainerDied","Data":"90066b4e910f0ee21a175275171e35716a5459a2c723385e352a837980db5011"} Dec 01 09:03:01 crc kubenswrapper[4873]: I1201 09:03:01.642670 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" event={"ID":"fef7b114-0e07-402d-a37b-315c36011f4b","Type":"ContainerStarted","Data":"dd1fd5c773938eaca06d208e8759b6376ff65c7cad3dc995d4b7c31e75f098dc"} Dec 01 09:03:01 crc kubenswrapper[4873]: I1201 09:03:01.642711 4873 scope.go:117] "RemoveContainer" containerID="0976ae6cae359bbc0353d195c023558e0d48d039655f0dbf281e76be957bb9b6" Dec 01 09:03:03 crc kubenswrapper[4873]: I1201 09:03:03.609899 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 01 09:03:03 crc kubenswrapper[4873]: I1201 09:03:03.612660 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 01 09:03:03 crc kubenswrapper[4873]: I1201 09:03:03.616626 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 01 09:03:03 crc kubenswrapper[4873]: I1201 09:03:03.692697 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 01 09:03:11 crc kubenswrapper[4873]: I1201 09:03:11.882599 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 09:03:12 crc kubenswrapper[4873]: I1201 09:03:12.791163 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 09:03:16 crc kubenswrapper[4873]: I1201 09:03:16.722723 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="6d27dec2-61e5-44ae-81df-8fee227ce3d8" containerName="rabbitmq" containerID="cri-o://59527d362e9a669ae5d6a07c74aacb56d24cf16f7963e8390b922ac5ab0c72b8" gracePeriod=604796 Dec 01 09:03:17 crc kubenswrapper[4873]: I1201 09:03:17.958501 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="29423166-c350-44f8-97ff-adb8b2b99165" containerName="rabbitmq" containerID="cri-o://74d71a8b0d6ed6192d61a25abdf207c390202505626ac490b22be2b31b6322a4" gracePeriod=604795 Dec 01 09:03:23 crc kubenswrapper[4873]: I1201 09:03:23.340374 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 01 09:03:23 crc kubenswrapper[4873]: I1201 09:03:23.481784 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6d27dec2-61e5-44ae-81df-8fee227ce3d8-rabbitmq-plugins\") pod \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\" (UID: \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\") " Dec 01 09:03:23 crc kubenswrapper[4873]: I1201 09:03:23.481954 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6d27dec2-61e5-44ae-81df-8fee227ce3d8-plugins-conf\") pod \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\" (UID: \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\") " Dec 01 09:03:23 crc kubenswrapper[4873]: I1201 09:03:23.482031 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\" (UID: \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\") " Dec 01 09:03:23 crc kubenswrapper[4873]: I1201 09:03:23.482077 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6d27dec2-61e5-44ae-81df-8fee227ce3d8-server-conf\") pod \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\" (UID: \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\") " Dec 01 09:03:23 crc kubenswrapper[4873]: I1201 09:03:23.482108 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6d27dec2-61e5-44ae-81df-8fee227ce3d8-rabbitmq-tls\") pod \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\" (UID: \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\") " Dec 01 09:03:23 crc kubenswrapper[4873]: I1201 09:03:23.482138 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6d27dec2-61e5-44ae-81df-8fee227ce3d8-erlang-cookie-secret\") pod \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\" (UID: \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\") " Dec 01 09:03:23 crc kubenswrapper[4873]: I1201 09:03:23.482176 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6d27dec2-61e5-44ae-81df-8fee227ce3d8-rabbitmq-confd\") pod \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\" (UID: \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\") " Dec 01 09:03:23 crc kubenswrapper[4873]: I1201 09:03:23.482203 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-96hg6\" (UniqueName: \"kubernetes.io/projected/6d27dec2-61e5-44ae-81df-8fee227ce3d8-kube-api-access-96hg6\") pod \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\" (UID: \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\") " Dec 01 09:03:23 crc kubenswrapper[4873]: I1201 09:03:23.482226 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6d27dec2-61e5-44ae-81df-8fee227ce3d8-config-data\") pod \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\" (UID: \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\") " Dec 01 09:03:23 crc kubenswrapper[4873]: I1201 09:03:23.482270 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6d27dec2-61e5-44ae-81df-8fee227ce3d8-pod-info\") pod \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\" (UID: \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\") " Dec 01 09:03:23 crc kubenswrapper[4873]: I1201 09:03:23.482325 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6d27dec2-61e5-44ae-81df-8fee227ce3d8-rabbitmq-erlang-cookie\") pod \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\" (UID: \"6d27dec2-61e5-44ae-81df-8fee227ce3d8\") " Dec 01 09:03:23 crc kubenswrapper[4873]: I1201 09:03:23.482797 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d27dec2-61e5-44ae-81df-8fee227ce3d8-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "6d27dec2-61e5-44ae-81df-8fee227ce3d8" (UID: "6d27dec2-61e5-44ae-81df-8fee227ce3d8"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:03:23 crc kubenswrapper[4873]: I1201 09:03:23.482903 4873 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6d27dec2-61e5-44ae-81df-8fee227ce3d8-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:23 crc kubenswrapper[4873]: I1201 09:03:23.483290 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d27dec2-61e5-44ae-81df-8fee227ce3d8-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "6d27dec2-61e5-44ae-81df-8fee227ce3d8" (UID: "6d27dec2-61e5-44ae-81df-8fee227ce3d8"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:03:23 crc kubenswrapper[4873]: I1201 09:03:23.488452 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d27dec2-61e5-44ae-81df-8fee227ce3d8-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "6d27dec2-61e5-44ae-81df-8fee227ce3d8" (UID: "6d27dec2-61e5-44ae-81df-8fee227ce3d8"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:03:23 crc kubenswrapper[4873]: I1201 09:03:23.502463 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d27dec2-61e5-44ae-81df-8fee227ce3d8-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "6d27dec2-61e5-44ae-81df-8fee227ce3d8" (UID: "6d27dec2-61e5-44ae-81df-8fee227ce3d8"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:03:23 crc kubenswrapper[4873]: I1201 09:03:23.502826 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "persistence") pod "6d27dec2-61e5-44ae-81df-8fee227ce3d8" (UID: "6d27dec2-61e5-44ae-81df-8fee227ce3d8"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:03:23 crc kubenswrapper[4873]: I1201 09:03:23.516335 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d27dec2-61e5-44ae-81df-8fee227ce3d8-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "6d27dec2-61e5-44ae-81df-8fee227ce3d8" (UID: "6d27dec2-61e5-44ae-81df-8fee227ce3d8"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:03:23 crc kubenswrapper[4873]: I1201 09:03:23.517595 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d27dec2-61e5-44ae-81df-8fee227ce3d8-kube-api-access-96hg6" (OuterVolumeSpecName: "kube-api-access-96hg6") pod "6d27dec2-61e5-44ae-81df-8fee227ce3d8" (UID: "6d27dec2-61e5-44ae-81df-8fee227ce3d8"). InnerVolumeSpecName "kube-api-access-96hg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:03:23 crc kubenswrapper[4873]: I1201 09:03:23.523647 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/6d27dec2-61e5-44ae-81df-8fee227ce3d8-pod-info" (OuterVolumeSpecName: "pod-info") pod "6d27dec2-61e5-44ae-81df-8fee227ce3d8" (UID: "6d27dec2-61e5-44ae-81df-8fee227ce3d8"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 01 09:03:23 crc kubenswrapper[4873]: I1201 09:03:23.534741 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d27dec2-61e5-44ae-81df-8fee227ce3d8-config-data" (OuterVolumeSpecName: "config-data") pod "6d27dec2-61e5-44ae-81df-8fee227ce3d8" (UID: "6d27dec2-61e5-44ae-81df-8fee227ce3d8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:03:23 crc kubenswrapper[4873]: I1201 09:03:23.544794 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d27dec2-61e5-44ae-81df-8fee227ce3d8-server-conf" (OuterVolumeSpecName: "server-conf") pod "6d27dec2-61e5-44ae-81df-8fee227ce3d8" (UID: "6d27dec2-61e5-44ae-81df-8fee227ce3d8"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:03:23 crc kubenswrapper[4873]: I1201 09:03:23.585603 4873 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6d27dec2-61e5-44ae-81df-8fee227ce3d8-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:23 crc kubenswrapper[4873]: I1201 09:03:23.585714 4873 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Dec 01 09:03:23 crc kubenswrapper[4873]: I1201 09:03:23.585736 4873 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6d27dec2-61e5-44ae-81df-8fee227ce3d8-server-conf\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:23 crc kubenswrapper[4873]: I1201 09:03:23.585751 4873 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6d27dec2-61e5-44ae-81df-8fee227ce3d8-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:23 crc kubenswrapper[4873]: I1201 09:03:23.585765 4873 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6d27dec2-61e5-44ae-81df-8fee227ce3d8-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:23 crc kubenswrapper[4873]: I1201 09:03:23.585802 4873 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6d27dec2-61e5-44ae-81df-8fee227ce3d8-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:23 crc kubenswrapper[4873]: I1201 09:03:23.585817 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-96hg6\" (UniqueName: \"kubernetes.io/projected/6d27dec2-61e5-44ae-81df-8fee227ce3d8-kube-api-access-96hg6\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:23 crc kubenswrapper[4873]: I1201 09:03:23.585834 4873 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6d27dec2-61e5-44ae-81df-8fee227ce3d8-pod-info\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:23 crc kubenswrapper[4873]: I1201 09:03:23.585846 4873 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6d27dec2-61e5-44ae-81df-8fee227ce3d8-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:23 crc kubenswrapper[4873]: I1201 09:03:23.618375 4873 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Dec 01 09:03:23 crc kubenswrapper[4873]: I1201 09:03:23.626916 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d27dec2-61e5-44ae-81df-8fee227ce3d8-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "6d27dec2-61e5-44ae-81df-8fee227ce3d8" (UID: "6d27dec2-61e5-44ae-81df-8fee227ce3d8"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:03:23 crc kubenswrapper[4873]: I1201 09:03:23.687188 4873 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:23 crc kubenswrapper[4873]: I1201 09:03:23.687222 4873 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6d27dec2-61e5-44ae-81df-8fee227ce3d8-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:23 crc kubenswrapper[4873]: I1201 09:03:23.949383 4873 generic.go:334] "Generic (PLEG): container finished" podID="6d27dec2-61e5-44ae-81df-8fee227ce3d8" containerID="59527d362e9a669ae5d6a07c74aacb56d24cf16f7963e8390b922ac5ab0c72b8" exitCode=0 Dec 01 09:03:23 crc kubenswrapper[4873]: I1201 09:03:23.949471 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6d27dec2-61e5-44ae-81df-8fee227ce3d8","Type":"ContainerDied","Data":"59527d362e9a669ae5d6a07c74aacb56d24cf16f7963e8390b922ac5ab0c72b8"} Dec 01 09:03:23 crc kubenswrapper[4873]: I1201 09:03:23.949533 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6d27dec2-61e5-44ae-81df-8fee227ce3d8","Type":"ContainerDied","Data":"0019bd449173284d5a8f1175b4ae0829a6e3e5bf457e17b6e629cfd3709c7620"} Dec 01 09:03:23 crc kubenswrapper[4873]: I1201 09:03:23.949581 4873 scope.go:117] "RemoveContainer" containerID="59527d362e9a669ae5d6a07c74aacb56d24cf16f7963e8390b922ac5ab0c72b8" Dec 01 09:03:23 crc kubenswrapper[4873]: I1201 09:03:23.949710 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.022251 4873 scope.go:117] "RemoveContainer" containerID="cfec876654e058bc3231e41fa25327a07bbec96f72ff60c123f47775b7ad0e34" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.029187 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.041757 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.049590 4873 scope.go:117] "RemoveContainer" containerID="59527d362e9a669ae5d6a07c74aacb56d24cf16f7963e8390b922ac5ab0c72b8" Dec 01 09:03:24 crc kubenswrapper[4873]: E1201 09:03:24.050164 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59527d362e9a669ae5d6a07c74aacb56d24cf16f7963e8390b922ac5ab0c72b8\": container with ID starting with 59527d362e9a669ae5d6a07c74aacb56d24cf16f7963e8390b922ac5ab0c72b8 not found: ID does not exist" containerID="59527d362e9a669ae5d6a07c74aacb56d24cf16f7963e8390b922ac5ab0c72b8" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.050199 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59527d362e9a669ae5d6a07c74aacb56d24cf16f7963e8390b922ac5ab0c72b8"} err="failed to get container status \"59527d362e9a669ae5d6a07c74aacb56d24cf16f7963e8390b922ac5ab0c72b8\": rpc error: code = NotFound desc = could not find container \"59527d362e9a669ae5d6a07c74aacb56d24cf16f7963e8390b922ac5ab0c72b8\": container with ID starting with 59527d362e9a669ae5d6a07c74aacb56d24cf16f7963e8390b922ac5ab0c72b8 not found: ID does not exist" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.050224 4873 scope.go:117] "RemoveContainer" containerID="cfec876654e058bc3231e41fa25327a07bbec96f72ff60c123f47775b7ad0e34" Dec 01 09:03:24 crc kubenswrapper[4873]: E1201 09:03:24.050554 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cfec876654e058bc3231e41fa25327a07bbec96f72ff60c123f47775b7ad0e34\": container with ID starting with cfec876654e058bc3231e41fa25327a07bbec96f72ff60c123f47775b7ad0e34 not found: ID does not exist" containerID="cfec876654e058bc3231e41fa25327a07bbec96f72ff60c123f47775b7ad0e34" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.050573 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfec876654e058bc3231e41fa25327a07bbec96f72ff60c123f47775b7ad0e34"} err="failed to get container status \"cfec876654e058bc3231e41fa25327a07bbec96f72ff60c123f47775b7ad0e34\": rpc error: code = NotFound desc = could not find container \"cfec876654e058bc3231e41fa25327a07bbec96f72ff60c123f47775b7ad0e34\": container with ID starting with cfec876654e058bc3231e41fa25327a07bbec96f72ff60c123f47775b7ad0e34 not found: ID does not exist" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.068640 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 09:03:24 crc kubenswrapper[4873]: E1201 09:03:24.069142 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d27dec2-61e5-44ae-81df-8fee227ce3d8" containerName="rabbitmq" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.069156 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d27dec2-61e5-44ae-81df-8fee227ce3d8" containerName="rabbitmq" Dec 01 09:03:24 crc kubenswrapper[4873]: E1201 09:03:24.069189 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d27dec2-61e5-44ae-81df-8fee227ce3d8" containerName="setup-container" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.069195 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d27dec2-61e5-44ae-81df-8fee227ce3d8" containerName="setup-container" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.069381 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d27dec2-61e5-44ae-81df-8fee227ce3d8" containerName="rabbitmq" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.070741 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.074232 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.074383 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.074590 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.074699 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-td8xf" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.075687 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.075988 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.077274 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.086222 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.209617 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/84aac4b8-9ed2-40d2-81f1-eba3af6e46b0-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"84aac4b8-9ed2-40d2-81f1-eba3af6e46b0\") " pod="openstack/rabbitmq-server-0" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.209678 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/84aac4b8-9ed2-40d2-81f1-eba3af6e46b0-server-conf\") pod \"rabbitmq-server-0\" (UID: \"84aac4b8-9ed2-40d2-81f1-eba3af6e46b0\") " pod="openstack/rabbitmq-server-0" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.209729 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/84aac4b8-9ed2-40d2-81f1-eba3af6e46b0-pod-info\") pod \"rabbitmq-server-0\" (UID: \"84aac4b8-9ed2-40d2-81f1-eba3af6e46b0\") " pod="openstack/rabbitmq-server-0" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.209754 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/84aac4b8-9ed2-40d2-81f1-eba3af6e46b0-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"84aac4b8-9ed2-40d2-81f1-eba3af6e46b0\") " pod="openstack/rabbitmq-server-0" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.209799 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/84aac4b8-9ed2-40d2-81f1-eba3af6e46b0-config-data\") pod \"rabbitmq-server-0\" (UID: \"84aac4b8-9ed2-40d2-81f1-eba3af6e46b0\") " pod="openstack/rabbitmq-server-0" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.209839 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"84aac4b8-9ed2-40d2-81f1-eba3af6e46b0\") " pod="openstack/rabbitmq-server-0" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.209856 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/84aac4b8-9ed2-40d2-81f1-eba3af6e46b0-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"84aac4b8-9ed2-40d2-81f1-eba3af6e46b0\") " pod="openstack/rabbitmq-server-0" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.209920 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/84aac4b8-9ed2-40d2-81f1-eba3af6e46b0-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"84aac4b8-9ed2-40d2-81f1-eba3af6e46b0\") " pod="openstack/rabbitmq-server-0" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.209979 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhzh6\" (UniqueName: \"kubernetes.io/projected/84aac4b8-9ed2-40d2-81f1-eba3af6e46b0-kube-api-access-xhzh6\") pod \"rabbitmq-server-0\" (UID: \"84aac4b8-9ed2-40d2-81f1-eba3af6e46b0\") " pod="openstack/rabbitmq-server-0" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.210034 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/84aac4b8-9ed2-40d2-81f1-eba3af6e46b0-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"84aac4b8-9ed2-40d2-81f1-eba3af6e46b0\") " pod="openstack/rabbitmq-server-0" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.210058 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/84aac4b8-9ed2-40d2-81f1-eba3af6e46b0-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"84aac4b8-9ed2-40d2-81f1-eba3af6e46b0\") " pod="openstack/rabbitmq-server-0" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.312158 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/84aac4b8-9ed2-40d2-81f1-eba3af6e46b0-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"84aac4b8-9ed2-40d2-81f1-eba3af6e46b0\") " pod="openstack/rabbitmq-server-0" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.312245 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/84aac4b8-9ed2-40d2-81f1-eba3af6e46b0-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"84aac4b8-9ed2-40d2-81f1-eba3af6e46b0\") " pod="openstack/rabbitmq-server-0" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.312326 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/84aac4b8-9ed2-40d2-81f1-eba3af6e46b0-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"84aac4b8-9ed2-40d2-81f1-eba3af6e46b0\") " pod="openstack/rabbitmq-server-0" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.312369 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/84aac4b8-9ed2-40d2-81f1-eba3af6e46b0-server-conf\") pod \"rabbitmq-server-0\" (UID: \"84aac4b8-9ed2-40d2-81f1-eba3af6e46b0\") " pod="openstack/rabbitmq-server-0" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.312426 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/84aac4b8-9ed2-40d2-81f1-eba3af6e46b0-pod-info\") pod \"rabbitmq-server-0\" (UID: \"84aac4b8-9ed2-40d2-81f1-eba3af6e46b0\") " pod="openstack/rabbitmq-server-0" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.312461 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/84aac4b8-9ed2-40d2-81f1-eba3af6e46b0-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"84aac4b8-9ed2-40d2-81f1-eba3af6e46b0\") " pod="openstack/rabbitmq-server-0" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.312531 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/84aac4b8-9ed2-40d2-81f1-eba3af6e46b0-config-data\") pod \"rabbitmq-server-0\" (UID: \"84aac4b8-9ed2-40d2-81f1-eba3af6e46b0\") " pod="openstack/rabbitmq-server-0" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.312571 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"84aac4b8-9ed2-40d2-81f1-eba3af6e46b0\") " pod="openstack/rabbitmq-server-0" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.312602 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/84aac4b8-9ed2-40d2-81f1-eba3af6e46b0-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"84aac4b8-9ed2-40d2-81f1-eba3af6e46b0\") " pod="openstack/rabbitmq-server-0" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.312702 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/84aac4b8-9ed2-40d2-81f1-eba3af6e46b0-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"84aac4b8-9ed2-40d2-81f1-eba3af6e46b0\") " pod="openstack/rabbitmq-server-0" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.312797 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhzh6\" (UniqueName: \"kubernetes.io/projected/84aac4b8-9ed2-40d2-81f1-eba3af6e46b0-kube-api-access-xhzh6\") pod \"rabbitmq-server-0\" (UID: \"84aac4b8-9ed2-40d2-81f1-eba3af6e46b0\") " pod="openstack/rabbitmq-server-0" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.315403 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/84aac4b8-9ed2-40d2-81f1-eba3af6e46b0-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"84aac4b8-9ed2-40d2-81f1-eba3af6e46b0\") " pod="openstack/rabbitmq-server-0" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.316259 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/84aac4b8-9ed2-40d2-81f1-eba3af6e46b0-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"84aac4b8-9ed2-40d2-81f1-eba3af6e46b0\") " pod="openstack/rabbitmq-server-0" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.317453 4873 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"84aac4b8-9ed2-40d2-81f1-eba3af6e46b0\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-server-0" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.319118 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/84aac4b8-9ed2-40d2-81f1-eba3af6e46b0-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"84aac4b8-9ed2-40d2-81f1-eba3af6e46b0\") " pod="openstack/rabbitmq-server-0" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.322939 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/84aac4b8-9ed2-40d2-81f1-eba3af6e46b0-config-data\") pod \"rabbitmq-server-0\" (UID: \"84aac4b8-9ed2-40d2-81f1-eba3af6e46b0\") " pod="openstack/rabbitmq-server-0" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.325274 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/84aac4b8-9ed2-40d2-81f1-eba3af6e46b0-server-conf\") pod \"rabbitmq-server-0\" (UID: \"84aac4b8-9ed2-40d2-81f1-eba3af6e46b0\") " pod="openstack/rabbitmq-server-0" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.333948 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/84aac4b8-9ed2-40d2-81f1-eba3af6e46b0-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"84aac4b8-9ed2-40d2-81f1-eba3af6e46b0\") " pod="openstack/rabbitmq-server-0" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.333954 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/84aac4b8-9ed2-40d2-81f1-eba3af6e46b0-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"84aac4b8-9ed2-40d2-81f1-eba3af6e46b0\") " pod="openstack/rabbitmq-server-0" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.340240 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/84aac4b8-9ed2-40d2-81f1-eba3af6e46b0-pod-info\") pod \"rabbitmq-server-0\" (UID: \"84aac4b8-9ed2-40d2-81f1-eba3af6e46b0\") " pod="openstack/rabbitmq-server-0" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.345785 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/84aac4b8-9ed2-40d2-81f1-eba3af6e46b0-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"84aac4b8-9ed2-40d2-81f1-eba3af6e46b0\") " pod="openstack/rabbitmq-server-0" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.369863 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhzh6\" (UniqueName: \"kubernetes.io/projected/84aac4b8-9ed2-40d2-81f1-eba3af6e46b0-kube-api-access-xhzh6\") pod \"rabbitmq-server-0\" (UID: \"84aac4b8-9ed2-40d2-81f1-eba3af6e46b0\") " pod="openstack/rabbitmq-server-0" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.404854 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"84aac4b8-9ed2-40d2-81f1-eba3af6e46b0\") " pod="openstack/rabbitmq-server-0" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.426655 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-td8xf" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.434676 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.467991 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d27dec2-61e5-44ae-81df-8fee227ce3d8" path="/var/lib/kubelet/pods/6d27dec2-61e5-44ae-81df-8fee227ce3d8/volumes" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.520621 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.619099 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/29423166-c350-44f8-97ff-adb8b2b99165-rabbitmq-plugins\") pod \"29423166-c350-44f8-97ff-adb8b2b99165\" (UID: \"29423166-c350-44f8-97ff-adb8b2b99165\") " Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.619167 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/29423166-c350-44f8-97ff-adb8b2b99165-erlang-cookie-secret\") pod \"29423166-c350-44f8-97ff-adb8b2b99165\" (UID: \"29423166-c350-44f8-97ff-adb8b2b99165\") " Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.619222 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/29423166-c350-44f8-97ff-adb8b2b99165-rabbitmq-confd\") pod \"29423166-c350-44f8-97ff-adb8b2b99165\" (UID: \"29423166-c350-44f8-97ff-adb8b2b99165\") " Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.619274 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/29423166-c350-44f8-97ff-adb8b2b99165-server-conf\") pod \"29423166-c350-44f8-97ff-adb8b2b99165\" (UID: \"29423166-c350-44f8-97ff-adb8b2b99165\") " Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.619303 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/29423166-c350-44f8-97ff-adb8b2b99165-rabbitmq-tls\") pod \"29423166-c350-44f8-97ff-adb8b2b99165\" (UID: \"29423166-c350-44f8-97ff-adb8b2b99165\") " Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.619324 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/29423166-c350-44f8-97ff-adb8b2b99165-rabbitmq-erlang-cookie\") pod \"29423166-c350-44f8-97ff-adb8b2b99165\" (UID: \"29423166-c350-44f8-97ff-adb8b2b99165\") " Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.619364 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/29423166-c350-44f8-97ff-adb8b2b99165-plugins-conf\") pod \"29423166-c350-44f8-97ff-adb8b2b99165\" (UID: \"29423166-c350-44f8-97ff-adb8b2b99165\") " Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.619406 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"29423166-c350-44f8-97ff-adb8b2b99165\" (UID: \"29423166-c350-44f8-97ff-adb8b2b99165\") " Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.619426 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/29423166-c350-44f8-97ff-adb8b2b99165-pod-info\") pod \"29423166-c350-44f8-97ff-adb8b2b99165\" (UID: \"29423166-c350-44f8-97ff-adb8b2b99165\") " Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.619456 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/29423166-c350-44f8-97ff-adb8b2b99165-config-data\") pod \"29423166-c350-44f8-97ff-adb8b2b99165\" (UID: \"29423166-c350-44f8-97ff-adb8b2b99165\") " Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.619540 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7qzmb\" (UniqueName: \"kubernetes.io/projected/29423166-c350-44f8-97ff-adb8b2b99165-kube-api-access-7qzmb\") pod \"29423166-c350-44f8-97ff-adb8b2b99165\" (UID: \"29423166-c350-44f8-97ff-adb8b2b99165\") " Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.625424 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29423166-c350-44f8-97ff-adb8b2b99165-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "29423166-c350-44f8-97ff-adb8b2b99165" (UID: "29423166-c350-44f8-97ff-adb8b2b99165"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.625451 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29423166-c350-44f8-97ff-adb8b2b99165-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "29423166-c350-44f8-97ff-adb8b2b99165" (UID: "29423166-c350-44f8-97ff-adb8b2b99165"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.633089 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29423166-c350-44f8-97ff-adb8b2b99165-kube-api-access-7qzmb" (OuterVolumeSpecName: "kube-api-access-7qzmb") pod "29423166-c350-44f8-97ff-adb8b2b99165" (UID: "29423166-c350-44f8-97ff-adb8b2b99165"). InnerVolumeSpecName "kube-api-access-7qzmb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.633099 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/29423166-c350-44f8-97ff-adb8b2b99165-pod-info" (OuterVolumeSpecName: "pod-info") pod "29423166-c350-44f8-97ff-adb8b2b99165" (UID: "29423166-c350-44f8-97ff-adb8b2b99165"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.633208 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "persistence") pod "29423166-c350-44f8-97ff-adb8b2b99165" (UID: "29423166-c350-44f8-97ff-adb8b2b99165"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.633433 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29423166-c350-44f8-97ff-adb8b2b99165-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "29423166-c350-44f8-97ff-adb8b2b99165" (UID: "29423166-c350-44f8-97ff-adb8b2b99165"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.637647 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/29423166-c350-44f8-97ff-adb8b2b99165-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "29423166-c350-44f8-97ff-adb8b2b99165" (UID: "29423166-c350-44f8-97ff-adb8b2b99165"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.644377 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29423166-c350-44f8-97ff-adb8b2b99165-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "29423166-c350-44f8-97ff-adb8b2b99165" (UID: "29423166-c350-44f8-97ff-adb8b2b99165"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.667562 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/29423166-c350-44f8-97ff-adb8b2b99165-config-data" (OuterVolumeSpecName: "config-data") pod "29423166-c350-44f8-97ff-adb8b2b99165" (UID: "29423166-c350-44f8-97ff-adb8b2b99165"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.716938 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/29423166-c350-44f8-97ff-adb8b2b99165-server-conf" (OuterVolumeSpecName: "server-conf") pod "29423166-c350-44f8-97ff-adb8b2b99165" (UID: "29423166-c350-44f8-97ff-adb8b2b99165"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.722893 4873 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/29423166-c350-44f8-97ff-adb8b2b99165-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.722942 4873 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/29423166-c350-44f8-97ff-adb8b2b99165-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.722957 4873 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/29423166-c350-44f8-97ff-adb8b2b99165-server-conf\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.722968 4873 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/29423166-c350-44f8-97ff-adb8b2b99165-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.722980 4873 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/29423166-c350-44f8-97ff-adb8b2b99165-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.722992 4873 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/29423166-c350-44f8-97ff-adb8b2b99165-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.723046 4873 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.723059 4873 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/29423166-c350-44f8-97ff-adb8b2b99165-pod-info\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.723069 4873 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/29423166-c350-44f8-97ff-adb8b2b99165-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.723082 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7qzmb\" (UniqueName: \"kubernetes.io/projected/29423166-c350-44f8-97ff-adb8b2b99165-kube-api-access-7qzmb\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.756624 4873 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.767380 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29423166-c350-44f8-97ff-adb8b2b99165-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "29423166-c350-44f8-97ff-adb8b2b99165" (UID: "29423166-c350-44f8-97ff-adb8b2b99165"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.827702 4873 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/29423166-c350-44f8-97ff-adb8b2b99165-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:24 crc kubenswrapper[4873]: I1201 09:03:24.827770 4873 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.031137 4873 generic.go:334] "Generic (PLEG): container finished" podID="29423166-c350-44f8-97ff-adb8b2b99165" containerID="74d71a8b0d6ed6192d61a25abdf207c390202505626ac490b22be2b31b6322a4" exitCode=0 Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.031182 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.031184 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"29423166-c350-44f8-97ff-adb8b2b99165","Type":"ContainerDied","Data":"74d71a8b0d6ed6192d61a25abdf207c390202505626ac490b22be2b31b6322a4"} Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.031958 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"29423166-c350-44f8-97ff-adb8b2b99165","Type":"ContainerDied","Data":"4db2a1c7ca11408f44768168a39319eac9b7db6c879e07052042c18efb0f7c98"} Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.031984 4873 scope.go:117] "RemoveContainer" containerID="74d71a8b0d6ed6192d61a25abdf207c390202505626ac490b22be2b31b6322a4" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.061891 4873 scope.go:117] "RemoveContainer" containerID="00ccca11cf2b2f7b96f7e096e15b6f797e1625129d5561faba84ede8dfacb4e9" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.078114 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.091535 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.105226 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.108930 4873 scope.go:117] "RemoveContainer" containerID="74d71a8b0d6ed6192d61a25abdf207c390202505626ac490b22be2b31b6322a4" Dec 01 09:03:25 crc kubenswrapper[4873]: E1201 09:03:25.116299 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74d71a8b0d6ed6192d61a25abdf207c390202505626ac490b22be2b31b6322a4\": container with ID starting with 74d71a8b0d6ed6192d61a25abdf207c390202505626ac490b22be2b31b6322a4 not found: ID does not exist" containerID="74d71a8b0d6ed6192d61a25abdf207c390202505626ac490b22be2b31b6322a4" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.116350 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74d71a8b0d6ed6192d61a25abdf207c390202505626ac490b22be2b31b6322a4"} err="failed to get container status \"74d71a8b0d6ed6192d61a25abdf207c390202505626ac490b22be2b31b6322a4\": rpc error: code = NotFound desc = could not find container \"74d71a8b0d6ed6192d61a25abdf207c390202505626ac490b22be2b31b6322a4\": container with ID starting with 74d71a8b0d6ed6192d61a25abdf207c390202505626ac490b22be2b31b6322a4 not found: ID does not exist" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.116386 4873 scope.go:117] "RemoveContainer" containerID="00ccca11cf2b2f7b96f7e096e15b6f797e1625129d5561faba84ede8dfacb4e9" Dec 01 09:03:25 crc kubenswrapper[4873]: E1201 09:03:25.116657 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00ccca11cf2b2f7b96f7e096e15b6f797e1625129d5561faba84ede8dfacb4e9\": container with ID starting with 00ccca11cf2b2f7b96f7e096e15b6f797e1625129d5561faba84ede8dfacb4e9 not found: ID does not exist" containerID="00ccca11cf2b2f7b96f7e096e15b6f797e1625129d5561faba84ede8dfacb4e9" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.116680 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00ccca11cf2b2f7b96f7e096e15b6f797e1625129d5561faba84ede8dfacb4e9"} err="failed to get container status \"00ccca11cf2b2f7b96f7e096e15b6f797e1625129d5561faba84ede8dfacb4e9\": rpc error: code = NotFound desc = could not find container \"00ccca11cf2b2f7b96f7e096e15b6f797e1625129d5561faba84ede8dfacb4e9\": container with ID starting with 00ccca11cf2b2f7b96f7e096e15b6f797e1625129d5561faba84ede8dfacb4e9 not found: ID does not exist" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.116789 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 09:03:25 crc kubenswrapper[4873]: E1201 09:03:25.117378 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29423166-c350-44f8-97ff-adb8b2b99165" containerName="rabbitmq" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.117403 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="29423166-c350-44f8-97ff-adb8b2b99165" containerName="rabbitmq" Dec 01 09:03:25 crc kubenswrapper[4873]: E1201 09:03:25.117431 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29423166-c350-44f8-97ff-adb8b2b99165" containerName="setup-container" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.117443 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="29423166-c350-44f8-97ff-adb8b2b99165" containerName="setup-container" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.117666 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="29423166-c350-44f8-97ff-adb8b2b99165" containerName="rabbitmq" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.118950 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.120482 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.122379 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.122983 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.123206 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.123443 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-zgbnc" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.123606 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.125060 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.189176 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.234759 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3baa8150-922b-4de7-ae44-24b9b5584212\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.234853 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tt4n\" (UniqueName: \"kubernetes.io/projected/3baa8150-922b-4de7-ae44-24b9b5584212-kube-api-access-5tt4n\") pod \"rabbitmq-cell1-server-0\" (UID: \"3baa8150-922b-4de7-ae44-24b9b5584212\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.234904 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3baa8150-922b-4de7-ae44-24b9b5584212-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3baa8150-922b-4de7-ae44-24b9b5584212\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.235075 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3baa8150-922b-4de7-ae44-24b9b5584212-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3baa8150-922b-4de7-ae44-24b9b5584212\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.235144 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3baa8150-922b-4de7-ae44-24b9b5584212-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3baa8150-922b-4de7-ae44-24b9b5584212\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.235314 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3baa8150-922b-4de7-ae44-24b9b5584212-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3baa8150-922b-4de7-ae44-24b9b5584212\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.235467 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3baa8150-922b-4de7-ae44-24b9b5584212-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3baa8150-922b-4de7-ae44-24b9b5584212\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.235533 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3baa8150-922b-4de7-ae44-24b9b5584212-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3baa8150-922b-4de7-ae44-24b9b5584212\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.235693 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3baa8150-922b-4de7-ae44-24b9b5584212-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3baa8150-922b-4de7-ae44-24b9b5584212\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.235809 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3baa8150-922b-4de7-ae44-24b9b5584212-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3baa8150-922b-4de7-ae44-24b9b5584212\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.235904 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3baa8150-922b-4de7-ae44-24b9b5584212-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3baa8150-922b-4de7-ae44-24b9b5584212\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.338131 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3baa8150-922b-4de7-ae44-24b9b5584212-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3baa8150-922b-4de7-ae44-24b9b5584212\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.338183 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3baa8150-922b-4de7-ae44-24b9b5584212-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3baa8150-922b-4de7-ae44-24b9b5584212\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.338216 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3baa8150-922b-4de7-ae44-24b9b5584212-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3baa8150-922b-4de7-ae44-24b9b5584212\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.338240 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3baa8150-922b-4de7-ae44-24b9b5584212-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3baa8150-922b-4de7-ae44-24b9b5584212\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.338261 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3baa8150-922b-4de7-ae44-24b9b5584212-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3baa8150-922b-4de7-ae44-24b9b5584212\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.338295 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3baa8150-922b-4de7-ae44-24b9b5584212\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.338332 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tt4n\" (UniqueName: \"kubernetes.io/projected/3baa8150-922b-4de7-ae44-24b9b5584212-kube-api-access-5tt4n\") pod \"rabbitmq-cell1-server-0\" (UID: \"3baa8150-922b-4de7-ae44-24b9b5584212\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.338364 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3baa8150-922b-4de7-ae44-24b9b5584212-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3baa8150-922b-4de7-ae44-24b9b5584212\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.338386 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3baa8150-922b-4de7-ae44-24b9b5584212-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3baa8150-922b-4de7-ae44-24b9b5584212\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.338403 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3baa8150-922b-4de7-ae44-24b9b5584212-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3baa8150-922b-4de7-ae44-24b9b5584212\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.338436 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3baa8150-922b-4de7-ae44-24b9b5584212-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3baa8150-922b-4de7-ae44-24b9b5584212\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.338931 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3baa8150-922b-4de7-ae44-24b9b5584212-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3baa8150-922b-4de7-ae44-24b9b5584212\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.339922 4873 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3baa8150-922b-4de7-ae44-24b9b5584212\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-cell1-server-0" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.340272 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3baa8150-922b-4de7-ae44-24b9b5584212-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3baa8150-922b-4de7-ae44-24b9b5584212\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.340752 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3baa8150-922b-4de7-ae44-24b9b5584212-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3baa8150-922b-4de7-ae44-24b9b5584212\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.341726 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3baa8150-922b-4de7-ae44-24b9b5584212-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3baa8150-922b-4de7-ae44-24b9b5584212\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.342347 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3baa8150-922b-4de7-ae44-24b9b5584212-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3baa8150-922b-4de7-ae44-24b9b5584212\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.345195 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3baa8150-922b-4de7-ae44-24b9b5584212-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3baa8150-922b-4de7-ae44-24b9b5584212\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.345955 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3baa8150-922b-4de7-ae44-24b9b5584212-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3baa8150-922b-4de7-ae44-24b9b5584212\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.346113 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3baa8150-922b-4de7-ae44-24b9b5584212-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3baa8150-922b-4de7-ae44-24b9b5584212\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.346709 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3baa8150-922b-4de7-ae44-24b9b5584212-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3baa8150-922b-4de7-ae44-24b9b5584212\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.370476 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tt4n\" (UniqueName: \"kubernetes.io/projected/3baa8150-922b-4de7-ae44-24b9b5584212-kube-api-access-5tt4n\") pod \"rabbitmq-cell1-server-0\" (UID: \"3baa8150-922b-4de7-ae44-24b9b5584212\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.383364 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3baa8150-922b-4de7-ae44-24b9b5584212\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 09:03:25 crc kubenswrapper[4873]: I1201 09:03:25.546396 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 01 09:03:26 crc kubenswrapper[4873]: I1201 09:03:26.046641 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"84aac4b8-9ed2-40d2-81f1-eba3af6e46b0","Type":"ContainerStarted","Data":"73b13634ccc69a329017bf5cc0c638f5c6c38a1678e4009f7df6053756d197ba"} Dec 01 09:03:26 crc kubenswrapper[4873]: W1201 09:03:26.083804 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3baa8150_922b_4de7_ae44_24b9b5584212.slice/crio-65589c16b3f6f10f044fa92974cb3beff2737d4e72cd479d95f77256e1c7bd62 WatchSource:0}: Error finding container 65589c16b3f6f10f044fa92974cb3beff2737d4e72cd479d95f77256e1c7bd62: Status 404 returned error can't find the container with id 65589c16b3f6f10f044fa92974cb3beff2737d4e72cd479d95f77256e1c7bd62 Dec 01 09:03:26 crc kubenswrapper[4873]: I1201 09:03:26.084757 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 09:03:26 crc kubenswrapper[4873]: I1201 09:03:26.444710 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29423166-c350-44f8-97ff-adb8b2b99165" path="/var/lib/kubelet/pods/29423166-c350-44f8-97ff-adb8b2b99165/volumes" Dec 01 09:03:27 crc kubenswrapper[4873]: I1201 09:03:27.068470 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3baa8150-922b-4de7-ae44-24b9b5584212","Type":"ContainerStarted","Data":"65589c16b3f6f10f044fa92974cb3beff2737d4e72cd479d95f77256e1c7bd62"} Dec 01 09:03:27 crc kubenswrapper[4873]: I1201 09:03:27.522486 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6447ccbd8f-rgfqr"] Dec 01 09:03:27 crc kubenswrapper[4873]: I1201 09:03:27.524799 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6447ccbd8f-rgfqr" Dec 01 09:03:27 crc kubenswrapper[4873]: I1201 09:03:27.531658 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 01 09:03:27 crc kubenswrapper[4873]: I1201 09:03:27.538157 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6447ccbd8f-rgfqr"] Dec 01 09:03:27 crc kubenswrapper[4873]: I1201 09:03:27.599467 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/3d7ff16e-2ecf-44de-9b10-5c60981d674c-openstack-edpm-ipam\") pod \"dnsmasq-dns-6447ccbd8f-rgfqr\" (UID: \"3d7ff16e-2ecf-44de-9b10-5c60981d674c\") " pod="openstack/dnsmasq-dns-6447ccbd8f-rgfqr" Dec 01 09:03:27 crc kubenswrapper[4873]: I1201 09:03:27.599545 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3d7ff16e-2ecf-44de-9b10-5c60981d674c-dns-svc\") pod \"dnsmasq-dns-6447ccbd8f-rgfqr\" (UID: \"3d7ff16e-2ecf-44de-9b10-5c60981d674c\") " pod="openstack/dnsmasq-dns-6447ccbd8f-rgfqr" Dec 01 09:03:27 crc kubenswrapper[4873]: I1201 09:03:27.599573 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d7ff16e-2ecf-44de-9b10-5c60981d674c-config\") pod \"dnsmasq-dns-6447ccbd8f-rgfqr\" (UID: \"3d7ff16e-2ecf-44de-9b10-5c60981d674c\") " pod="openstack/dnsmasq-dns-6447ccbd8f-rgfqr" Dec 01 09:03:27 crc kubenswrapper[4873]: I1201 09:03:27.599620 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3d7ff16e-2ecf-44de-9b10-5c60981d674c-ovsdbserver-nb\") pod \"dnsmasq-dns-6447ccbd8f-rgfqr\" (UID: \"3d7ff16e-2ecf-44de-9b10-5c60981d674c\") " pod="openstack/dnsmasq-dns-6447ccbd8f-rgfqr" Dec 01 09:03:27 crc kubenswrapper[4873]: I1201 09:03:27.599662 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q46ff\" (UniqueName: \"kubernetes.io/projected/3d7ff16e-2ecf-44de-9b10-5c60981d674c-kube-api-access-q46ff\") pod \"dnsmasq-dns-6447ccbd8f-rgfqr\" (UID: \"3d7ff16e-2ecf-44de-9b10-5c60981d674c\") " pod="openstack/dnsmasq-dns-6447ccbd8f-rgfqr" Dec 01 09:03:27 crc kubenswrapper[4873]: I1201 09:03:27.599774 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3d7ff16e-2ecf-44de-9b10-5c60981d674c-ovsdbserver-sb\") pod \"dnsmasq-dns-6447ccbd8f-rgfqr\" (UID: \"3d7ff16e-2ecf-44de-9b10-5c60981d674c\") " pod="openstack/dnsmasq-dns-6447ccbd8f-rgfqr" Dec 01 09:03:27 crc kubenswrapper[4873]: I1201 09:03:27.701468 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3d7ff16e-2ecf-44de-9b10-5c60981d674c-dns-svc\") pod \"dnsmasq-dns-6447ccbd8f-rgfqr\" (UID: \"3d7ff16e-2ecf-44de-9b10-5c60981d674c\") " pod="openstack/dnsmasq-dns-6447ccbd8f-rgfqr" Dec 01 09:03:27 crc kubenswrapper[4873]: I1201 09:03:27.701533 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d7ff16e-2ecf-44de-9b10-5c60981d674c-config\") pod \"dnsmasq-dns-6447ccbd8f-rgfqr\" (UID: \"3d7ff16e-2ecf-44de-9b10-5c60981d674c\") " pod="openstack/dnsmasq-dns-6447ccbd8f-rgfqr" Dec 01 09:03:27 crc kubenswrapper[4873]: I1201 09:03:27.701593 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3d7ff16e-2ecf-44de-9b10-5c60981d674c-ovsdbserver-nb\") pod \"dnsmasq-dns-6447ccbd8f-rgfqr\" (UID: \"3d7ff16e-2ecf-44de-9b10-5c60981d674c\") " pod="openstack/dnsmasq-dns-6447ccbd8f-rgfqr" Dec 01 09:03:27 crc kubenswrapper[4873]: I1201 09:03:27.701624 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q46ff\" (UniqueName: \"kubernetes.io/projected/3d7ff16e-2ecf-44de-9b10-5c60981d674c-kube-api-access-q46ff\") pod \"dnsmasq-dns-6447ccbd8f-rgfqr\" (UID: \"3d7ff16e-2ecf-44de-9b10-5c60981d674c\") " pod="openstack/dnsmasq-dns-6447ccbd8f-rgfqr" Dec 01 09:03:27 crc kubenswrapper[4873]: I1201 09:03:27.701697 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3d7ff16e-2ecf-44de-9b10-5c60981d674c-ovsdbserver-sb\") pod \"dnsmasq-dns-6447ccbd8f-rgfqr\" (UID: \"3d7ff16e-2ecf-44de-9b10-5c60981d674c\") " pod="openstack/dnsmasq-dns-6447ccbd8f-rgfqr" Dec 01 09:03:27 crc kubenswrapper[4873]: I1201 09:03:27.701731 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/3d7ff16e-2ecf-44de-9b10-5c60981d674c-openstack-edpm-ipam\") pod \"dnsmasq-dns-6447ccbd8f-rgfqr\" (UID: \"3d7ff16e-2ecf-44de-9b10-5c60981d674c\") " pod="openstack/dnsmasq-dns-6447ccbd8f-rgfqr" Dec 01 09:03:27 crc kubenswrapper[4873]: I1201 09:03:27.702976 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3d7ff16e-2ecf-44de-9b10-5c60981d674c-ovsdbserver-sb\") pod \"dnsmasq-dns-6447ccbd8f-rgfqr\" (UID: \"3d7ff16e-2ecf-44de-9b10-5c60981d674c\") " pod="openstack/dnsmasq-dns-6447ccbd8f-rgfqr" Dec 01 09:03:27 crc kubenswrapper[4873]: I1201 09:03:27.703086 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3d7ff16e-2ecf-44de-9b10-5c60981d674c-dns-svc\") pod \"dnsmasq-dns-6447ccbd8f-rgfqr\" (UID: \"3d7ff16e-2ecf-44de-9b10-5c60981d674c\") " pod="openstack/dnsmasq-dns-6447ccbd8f-rgfqr" Dec 01 09:03:27 crc kubenswrapper[4873]: I1201 09:03:27.703141 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/3d7ff16e-2ecf-44de-9b10-5c60981d674c-openstack-edpm-ipam\") pod \"dnsmasq-dns-6447ccbd8f-rgfqr\" (UID: \"3d7ff16e-2ecf-44de-9b10-5c60981d674c\") " pod="openstack/dnsmasq-dns-6447ccbd8f-rgfqr" Dec 01 09:03:27 crc kubenswrapper[4873]: I1201 09:03:27.703205 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3d7ff16e-2ecf-44de-9b10-5c60981d674c-ovsdbserver-nb\") pod \"dnsmasq-dns-6447ccbd8f-rgfqr\" (UID: \"3d7ff16e-2ecf-44de-9b10-5c60981d674c\") " pod="openstack/dnsmasq-dns-6447ccbd8f-rgfqr" Dec 01 09:03:27 crc kubenswrapper[4873]: I1201 09:03:27.703648 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d7ff16e-2ecf-44de-9b10-5c60981d674c-config\") pod \"dnsmasq-dns-6447ccbd8f-rgfqr\" (UID: \"3d7ff16e-2ecf-44de-9b10-5c60981d674c\") " pod="openstack/dnsmasq-dns-6447ccbd8f-rgfqr" Dec 01 09:03:27 crc kubenswrapper[4873]: I1201 09:03:27.741578 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q46ff\" (UniqueName: \"kubernetes.io/projected/3d7ff16e-2ecf-44de-9b10-5c60981d674c-kube-api-access-q46ff\") pod \"dnsmasq-dns-6447ccbd8f-rgfqr\" (UID: \"3d7ff16e-2ecf-44de-9b10-5c60981d674c\") " pod="openstack/dnsmasq-dns-6447ccbd8f-rgfqr" Dec 01 09:03:27 crc kubenswrapper[4873]: I1201 09:03:27.880852 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6447ccbd8f-rgfqr" Dec 01 09:03:28 crc kubenswrapper[4873]: I1201 09:03:28.108466 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"84aac4b8-9ed2-40d2-81f1-eba3af6e46b0","Type":"ContainerStarted","Data":"f6ecd62b7b64755500accd3a857b33968f234e31b8cb9d782b3e3626cf2fd48a"} Dec 01 09:03:28 crc kubenswrapper[4873]: I1201 09:03:28.426906 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6447ccbd8f-rgfqr"] Dec 01 09:03:28 crc kubenswrapper[4873]: W1201 09:03:28.435138 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3d7ff16e_2ecf_44de_9b10_5c60981d674c.slice/crio-c8aef672ba09680bc8ac2afcfd4102a5b972bc9bfdf8e83a1a07c5ab02f994a4 WatchSource:0}: Error finding container c8aef672ba09680bc8ac2afcfd4102a5b972bc9bfdf8e83a1a07c5ab02f994a4: Status 404 returned error can't find the container with id c8aef672ba09680bc8ac2afcfd4102a5b972bc9bfdf8e83a1a07c5ab02f994a4 Dec 01 09:03:29 crc kubenswrapper[4873]: I1201 09:03:29.120360 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3baa8150-922b-4de7-ae44-24b9b5584212","Type":"ContainerStarted","Data":"d6285fe2b3b994bd13a0add987955acea8a299727a6960e9c0a06224b4764e50"} Dec 01 09:03:29 crc kubenswrapper[4873]: I1201 09:03:29.123915 4873 generic.go:334] "Generic (PLEG): container finished" podID="3d7ff16e-2ecf-44de-9b10-5c60981d674c" containerID="ba0b6bc5473f347d003e0106e0fbe6b2edc21402e8b91ad650e4348efa779653" exitCode=0 Dec 01 09:03:29 crc kubenswrapper[4873]: I1201 09:03:29.124123 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6447ccbd8f-rgfqr" event={"ID":"3d7ff16e-2ecf-44de-9b10-5c60981d674c","Type":"ContainerDied","Data":"ba0b6bc5473f347d003e0106e0fbe6b2edc21402e8b91ad650e4348efa779653"} Dec 01 09:03:29 crc kubenswrapper[4873]: I1201 09:03:29.124167 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6447ccbd8f-rgfqr" event={"ID":"3d7ff16e-2ecf-44de-9b10-5c60981d674c","Type":"ContainerStarted","Data":"c8aef672ba09680bc8ac2afcfd4102a5b972bc9bfdf8e83a1a07c5ab02f994a4"} Dec 01 09:03:30 crc kubenswrapper[4873]: I1201 09:03:30.142362 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6447ccbd8f-rgfqr" event={"ID":"3d7ff16e-2ecf-44de-9b10-5c60981d674c","Type":"ContainerStarted","Data":"da46494a66f06ddde14ed0184bcadb65bcb1085b8fa27f8069d81792116386b1"} Dec 01 09:03:30 crc kubenswrapper[4873]: I1201 09:03:30.182171 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6447ccbd8f-rgfqr" podStartSLOduration=3.182140653 podStartE2EDuration="3.182140653s" podCreationTimestamp="2025-12-01 09:03:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:03:30.17016262 +0000 UTC m=+1386.072271199" watchObservedRunningTime="2025-12-01 09:03:30.182140653 +0000 UTC m=+1386.084249232" Dec 01 09:03:31 crc kubenswrapper[4873]: I1201 09:03:31.156058 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6447ccbd8f-rgfqr" Dec 01 09:03:37 crc kubenswrapper[4873]: I1201 09:03:37.910383 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6447ccbd8f-rgfqr" Dec 01 09:03:37 crc kubenswrapper[4873]: I1201 09:03:37.999599 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b856c5697-rw4vp"] Dec 01 09:03:37 crc kubenswrapper[4873]: I1201 09:03:37.999889 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b856c5697-rw4vp" podUID="e825def4-1364-4a5b-8896-b8b633a8142c" containerName="dnsmasq-dns" containerID="cri-o://05c253da90588e1e95a9fed1a64c169ae4c1cfbcbeaf1410c5a179ba9d783548" gracePeriod=10 Dec 01 09:03:38 crc kubenswrapper[4873]: I1201 09:03:38.168447 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-864d5fc68c-r9f7f"] Dec 01 09:03:38 crc kubenswrapper[4873]: I1201 09:03:38.170803 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-864d5fc68c-r9f7f" Dec 01 09:03:38 crc kubenswrapper[4873]: I1201 09:03:38.195268 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-864d5fc68c-r9f7f"] Dec 01 09:03:38 crc kubenswrapper[4873]: I1201 09:03:38.256576 4873 generic.go:334] "Generic (PLEG): container finished" podID="e825def4-1364-4a5b-8896-b8b633a8142c" containerID="05c253da90588e1e95a9fed1a64c169ae4c1cfbcbeaf1410c5a179ba9d783548" exitCode=0 Dec 01 09:03:38 crc kubenswrapper[4873]: I1201 09:03:38.256641 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b856c5697-rw4vp" event={"ID":"e825def4-1364-4a5b-8896-b8b633a8142c","Type":"ContainerDied","Data":"05c253da90588e1e95a9fed1a64c169ae4c1cfbcbeaf1410c5a179ba9d783548"} Dec 01 09:03:38 crc kubenswrapper[4873]: I1201 09:03:38.350281 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8d7e3564-5684-4792-b3b0-57001892f591-dns-svc\") pod \"dnsmasq-dns-864d5fc68c-r9f7f\" (UID: \"8d7e3564-5684-4792-b3b0-57001892f591\") " pod="openstack/dnsmasq-dns-864d5fc68c-r9f7f" Dec 01 09:03:38 crc kubenswrapper[4873]: I1201 09:03:38.350802 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/8d7e3564-5684-4792-b3b0-57001892f591-openstack-edpm-ipam\") pod \"dnsmasq-dns-864d5fc68c-r9f7f\" (UID: \"8d7e3564-5684-4792-b3b0-57001892f591\") " pod="openstack/dnsmasq-dns-864d5fc68c-r9f7f" Dec 01 09:03:38 crc kubenswrapper[4873]: I1201 09:03:38.350828 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcc8x\" (UniqueName: \"kubernetes.io/projected/8d7e3564-5684-4792-b3b0-57001892f591-kube-api-access-vcc8x\") pod \"dnsmasq-dns-864d5fc68c-r9f7f\" (UID: \"8d7e3564-5684-4792-b3b0-57001892f591\") " pod="openstack/dnsmasq-dns-864d5fc68c-r9f7f" Dec 01 09:03:38 crc kubenswrapper[4873]: I1201 09:03:38.350992 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8d7e3564-5684-4792-b3b0-57001892f591-ovsdbserver-nb\") pod \"dnsmasq-dns-864d5fc68c-r9f7f\" (UID: \"8d7e3564-5684-4792-b3b0-57001892f591\") " pod="openstack/dnsmasq-dns-864d5fc68c-r9f7f" Dec 01 09:03:38 crc kubenswrapper[4873]: I1201 09:03:38.351240 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8d7e3564-5684-4792-b3b0-57001892f591-ovsdbserver-sb\") pod \"dnsmasq-dns-864d5fc68c-r9f7f\" (UID: \"8d7e3564-5684-4792-b3b0-57001892f591\") " pod="openstack/dnsmasq-dns-864d5fc68c-r9f7f" Dec 01 09:03:38 crc kubenswrapper[4873]: I1201 09:03:38.351298 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d7e3564-5684-4792-b3b0-57001892f591-config\") pod \"dnsmasq-dns-864d5fc68c-r9f7f\" (UID: \"8d7e3564-5684-4792-b3b0-57001892f591\") " pod="openstack/dnsmasq-dns-864d5fc68c-r9f7f" Dec 01 09:03:38 crc kubenswrapper[4873]: I1201 09:03:38.452049 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8d7e3564-5684-4792-b3b0-57001892f591-ovsdbserver-nb\") pod \"dnsmasq-dns-864d5fc68c-r9f7f\" (UID: \"8d7e3564-5684-4792-b3b0-57001892f591\") " pod="openstack/dnsmasq-dns-864d5fc68c-r9f7f" Dec 01 09:03:38 crc kubenswrapper[4873]: I1201 09:03:38.452121 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8d7e3564-5684-4792-b3b0-57001892f591-ovsdbserver-sb\") pod \"dnsmasq-dns-864d5fc68c-r9f7f\" (UID: \"8d7e3564-5684-4792-b3b0-57001892f591\") " pod="openstack/dnsmasq-dns-864d5fc68c-r9f7f" Dec 01 09:03:38 crc kubenswrapper[4873]: I1201 09:03:38.452148 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d7e3564-5684-4792-b3b0-57001892f591-config\") pod \"dnsmasq-dns-864d5fc68c-r9f7f\" (UID: \"8d7e3564-5684-4792-b3b0-57001892f591\") " pod="openstack/dnsmasq-dns-864d5fc68c-r9f7f" Dec 01 09:03:38 crc kubenswrapper[4873]: I1201 09:03:38.452237 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8d7e3564-5684-4792-b3b0-57001892f591-dns-svc\") pod \"dnsmasq-dns-864d5fc68c-r9f7f\" (UID: \"8d7e3564-5684-4792-b3b0-57001892f591\") " pod="openstack/dnsmasq-dns-864d5fc68c-r9f7f" Dec 01 09:03:38 crc kubenswrapper[4873]: I1201 09:03:38.452259 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/8d7e3564-5684-4792-b3b0-57001892f591-openstack-edpm-ipam\") pod \"dnsmasq-dns-864d5fc68c-r9f7f\" (UID: \"8d7e3564-5684-4792-b3b0-57001892f591\") " pod="openstack/dnsmasq-dns-864d5fc68c-r9f7f" Dec 01 09:03:38 crc kubenswrapper[4873]: I1201 09:03:38.452274 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcc8x\" (UniqueName: \"kubernetes.io/projected/8d7e3564-5684-4792-b3b0-57001892f591-kube-api-access-vcc8x\") pod \"dnsmasq-dns-864d5fc68c-r9f7f\" (UID: \"8d7e3564-5684-4792-b3b0-57001892f591\") " pod="openstack/dnsmasq-dns-864d5fc68c-r9f7f" Dec 01 09:03:38 crc kubenswrapper[4873]: I1201 09:03:38.453515 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8d7e3564-5684-4792-b3b0-57001892f591-ovsdbserver-nb\") pod \"dnsmasq-dns-864d5fc68c-r9f7f\" (UID: \"8d7e3564-5684-4792-b3b0-57001892f591\") " pod="openstack/dnsmasq-dns-864d5fc68c-r9f7f" Dec 01 09:03:38 crc kubenswrapper[4873]: I1201 09:03:38.453994 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8d7e3564-5684-4792-b3b0-57001892f591-ovsdbserver-sb\") pod \"dnsmasq-dns-864d5fc68c-r9f7f\" (UID: \"8d7e3564-5684-4792-b3b0-57001892f591\") " pod="openstack/dnsmasq-dns-864d5fc68c-r9f7f" Dec 01 09:03:38 crc kubenswrapper[4873]: I1201 09:03:38.454458 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8d7e3564-5684-4792-b3b0-57001892f591-dns-svc\") pod \"dnsmasq-dns-864d5fc68c-r9f7f\" (UID: \"8d7e3564-5684-4792-b3b0-57001892f591\") " pod="openstack/dnsmasq-dns-864d5fc68c-r9f7f" Dec 01 09:03:38 crc kubenswrapper[4873]: I1201 09:03:38.454685 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d7e3564-5684-4792-b3b0-57001892f591-config\") pod \"dnsmasq-dns-864d5fc68c-r9f7f\" (UID: \"8d7e3564-5684-4792-b3b0-57001892f591\") " pod="openstack/dnsmasq-dns-864d5fc68c-r9f7f" Dec 01 09:03:38 crc kubenswrapper[4873]: I1201 09:03:38.455005 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/8d7e3564-5684-4792-b3b0-57001892f591-openstack-edpm-ipam\") pod \"dnsmasq-dns-864d5fc68c-r9f7f\" (UID: \"8d7e3564-5684-4792-b3b0-57001892f591\") " pod="openstack/dnsmasq-dns-864d5fc68c-r9f7f" Dec 01 09:03:38 crc kubenswrapper[4873]: I1201 09:03:38.473893 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcc8x\" (UniqueName: \"kubernetes.io/projected/8d7e3564-5684-4792-b3b0-57001892f591-kube-api-access-vcc8x\") pod \"dnsmasq-dns-864d5fc68c-r9f7f\" (UID: \"8d7e3564-5684-4792-b3b0-57001892f591\") " pod="openstack/dnsmasq-dns-864d5fc68c-r9f7f" Dec 01 09:03:38 crc kubenswrapper[4873]: I1201 09:03:38.535648 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b856c5697-rw4vp" Dec 01 09:03:38 crc kubenswrapper[4873]: I1201 09:03:38.557474 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-864d5fc68c-r9f7f" Dec 01 09:03:38 crc kubenswrapper[4873]: I1201 09:03:38.557532 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85n65\" (UniqueName: \"kubernetes.io/projected/e825def4-1364-4a5b-8896-b8b633a8142c-kube-api-access-85n65\") pod \"e825def4-1364-4a5b-8896-b8b633a8142c\" (UID: \"e825def4-1364-4a5b-8896-b8b633a8142c\") " Dec 01 09:03:38 crc kubenswrapper[4873]: I1201 09:03:38.557675 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e825def4-1364-4a5b-8896-b8b633a8142c-ovsdbserver-nb\") pod \"e825def4-1364-4a5b-8896-b8b633a8142c\" (UID: \"e825def4-1364-4a5b-8896-b8b633a8142c\") " Dec 01 09:03:38 crc kubenswrapper[4873]: I1201 09:03:38.557780 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e825def4-1364-4a5b-8896-b8b633a8142c-ovsdbserver-sb\") pod \"e825def4-1364-4a5b-8896-b8b633a8142c\" (UID: \"e825def4-1364-4a5b-8896-b8b633a8142c\") " Dec 01 09:03:38 crc kubenswrapper[4873]: I1201 09:03:38.557835 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e825def4-1364-4a5b-8896-b8b633a8142c-dns-svc\") pod \"e825def4-1364-4a5b-8896-b8b633a8142c\" (UID: \"e825def4-1364-4a5b-8896-b8b633a8142c\") " Dec 01 09:03:38 crc kubenswrapper[4873]: I1201 09:03:38.557853 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e825def4-1364-4a5b-8896-b8b633a8142c-config\") pod \"e825def4-1364-4a5b-8896-b8b633a8142c\" (UID: \"e825def4-1364-4a5b-8896-b8b633a8142c\") " Dec 01 09:03:38 crc kubenswrapper[4873]: I1201 09:03:38.563413 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e825def4-1364-4a5b-8896-b8b633a8142c-kube-api-access-85n65" (OuterVolumeSpecName: "kube-api-access-85n65") pod "e825def4-1364-4a5b-8896-b8b633a8142c" (UID: "e825def4-1364-4a5b-8896-b8b633a8142c"). InnerVolumeSpecName "kube-api-access-85n65". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:03:38 crc kubenswrapper[4873]: I1201 09:03:38.621984 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e825def4-1364-4a5b-8896-b8b633a8142c-config" (OuterVolumeSpecName: "config") pod "e825def4-1364-4a5b-8896-b8b633a8142c" (UID: "e825def4-1364-4a5b-8896-b8b633a8142c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:03:38 crc kubenswrapper[4873]: I1201 09:03:38.623008 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e825def4-1364-4a5b-8896-b8b633a8142c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e825def4-1364-4a5b-8896-b8b633a8142c" (UID: "e825def4-1364-4a5b-8896-b8b633a8142c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:03:38 crc kubenswrapper[4873]: I1201 09:03:38.643335 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e825def4-1364-4a5b-8896-b8b633a8142c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e825def4-1364-4a5b-8896-b8b633a8142c" (UID: "e825def4-1364-4a5b-8896-b8b633a8142c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:03:38 crc kubenswrapper[4873]: I1201 09:03:38.645499 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e825def4-1364-4a5b-8896-b8b633a8142c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e825def4-1364-4a5b-8896-b8b633a8142c" (UID: "e825def4-1364-4a5b-8896-b8b633a8142c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:03:38 crc kubenswrapper[4873]: I1201 09:03:38.660748 4873 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e825def4-1364-4a5b-8896-b8b633a8142c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:38 crc kubenswrapper[4873]: I1201 09:03:38.660772 4873 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e825def4-1364-4a5b-8896-b8b633a8142c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:38 crc kubenswrapper[4873]: I1201 09:03:38.660783 4873 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e825def4-1364-4a5b-8896-b8b633a8142c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:38 crc kubenswrapper[4873]: I1201 09:03:38.660793 4873 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e825def4-1364-4a5b-8896-b8b633a8142c-config\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:38 crc kubenswrapper[4873]: I1201 09:03:38.660806 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85n65\" (UniqueName: \"kubernetes.io/projected/e825def4-1364-4a5b-8896-b8b633a8142c-kube-api-access-85n65\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:39 crc kubenswrapper[4873]: I1201 09:03:39.139665 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-864d5fc68c-r9f7f"] Dec 01 09:03:39 crc kubenswrapper[4873]: W1201 09:03:39.150585 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8d7e3564_5684_4792_b3b0_57001892f591.slice/crio-63c28c34179fa261df3a34e7fcbb87902a2b62dab87592559d5813b491e11307 WatchSource:0}: Error finding container 63c28c34179fa261df3a34e7fcbb87902a2b62dab87592559d5813b491e11307: Status 404 returned error can't find the container with id 63c28c34179fa261df3a34e7fcbb87902a2b62dab87592559d5813b491e11307 Dec 01 09:03:39 crc kubenswrapper[4873]: I1201 09:03:39.280344 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b856c5697-rw4vp" Dec 01 09:03:39 crc kubenswrapper[4873]: I1201 09:03:39.280395 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b856c5697-rw4vp" event={"ID":"e825def4-1364-4a5b-8896-b8b633a8142c","Type":"ContainerDied","Data":"80f2df72089fa7908fc4b20d86ac3b58d77f85782af61ece0e62922200286d8d"} Dec 01 09:03:39 crc kubenswrapper[4873]: I1201 09:03:39.281076 4873 scope.go:117] "RemoveContainer" containerID="05c253da90588e1e95a9fed1a64c169ae4c1cfbcbeaf1410c5a179ba9d783548" Dec 01 09:03:39 crc kubenswrapper[4873]: I1201 09:03:39.284895 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-864d5fc68c-r9f7f" event={"ID":"8d7e3564-5684-4792-b3b0-57001892f591","Type":"ContainerStarted","Data":"63c28c34179fa261df3a34e7fcbb87902a2b62dab87592559d5813b491e11307"} Dec 01 09:03:39 crc kubenswrapper[4873]: I1201 09:03:39.347964 4873 scope.go:117] "RemoveContainer" containerID="d658bdfceb80bae9dd70080cb1de56f99de4162da33c63f48de1866ad682e7e7" Dec 01 09:03:39 crc kubenswrapper[4873]: I1201 09:03:39.349003 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b856c5697-rw4vp"] Dec 01 09:03:39 crc kubenswrapper[4873]: I1201 09:03:39.360738 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b856c5697-rw4vp"] Dec 01 09:03:40 crc kubenswrapper[4873]: I1201 09:03:40.299406 4873 generic.go:334] "Generic (PLEG): container finished" podID="8d7e3564-5684-4792-b3b0-57001892f591" containerID="f30c0d46b1aeaa6aedf219761bdb926b5662dc6e6cb3acf1194f97b216eeb0d1" exitCode=0 Dec 01 09:03:40 crc kubenswrapper[4873]: I1201 09:03:40.301477 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-864d5fc68c-r9f7f" event={"ID":"8d7e3564-5684-4792-b3b0-57001892f591","Type":"ContainerDied","Data":"f30c0d46b1aeaa6aedf219761bdb926b5662dc6e6cb3acf1194f97b216eeb0d1"} Dec 01 09:03:40 crc kubenswrapper[4873]: I1201 09:03:40.452342 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e825def4-1364-4a5b-8896-b8b633a8142c" path="/var/lib/kubelet/pods/e825def4-1364-4a5b-8896-b8b633a8142c/volumes" Dec 01 09:03:41 crc kubenswrapper[4873]: I1201 09:03:41.321400 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-864d5fc68c-r9f7f" event={"ID":"8d7e3564-5684-4792-b3b0-57001892f591","Type":"ContainerStarted","Data":"26f55973bcf4d0bcc3ac93c0eb1796034503b90c45a1929dc634866c66402344"} Dec 01 09:03:41 crc kubenswrapper[4873]: I1201 09:03:41.322453 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-864d5fc68c-r9f7f" Dec 01 09:03:48 crc kubenswrapper[4873]: I1201 09:03:48.560246 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-864d5fc68c-r9f7f" Dec 01 09:03:48 crc kubenswrapper[4873]: I1201 09:03:48.587404 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-864d5fc68c-r9f7f" podStartSLOduration=10.587383077 podStartE2EDuration="10.587383077s" podCreationTimestamp="2025-12-01 09:03:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:03:41.391312138 +0000 UTC m=+1397.293420687" watchObservedRunningTime="2025-12-01 09:03:48.587383077 +0000 UTC m=+1404.489491616" Dec 01 09:03:48 crc kubenswrapper[4873]: I1201 09:03:48.680670 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6447ccbd8f-rgfqr"] Dec 01 09:03:48 crc kubenswrapper[4873]: I1201 09:03:48.680910 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6447ccbd8f-rgfqr" podUID="3d7ff16e-2ecf-44de-9b10-5c60981d674c" containerName="dnsmasq-dns" containerID="cri-o://da46494a66f06ddde14ed0184bcadb65bcb1085b8fa27f8069d81792116386b1" gracePeriod=10 Dec 01 09:03:49 crc kubenswrapper[4873]: I1201 09:03:49.250349 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6447ccbd8f-rgfqr" Dec 01 09:03:49 crc kubenswrapper[4873]: I1201 09:03:49.394028 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3d7ff16e-2ecf-44de-9b10-5c60981d674c-ovsdbserver-sb\") pod \"3d7ff16e-2ecf-44de-9b10-5c60981d674c\" (UID: \"3d7ff16e-2ecf-44de-9b10-5c60981d674c\") " Dec 01 09:03:49 crc kubenswrapper[4873]: I1201 09:03:49.394139 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/3d7ff16e-2ecf-44de-9b10-5c60981d674c-openstack-edpm-ipam\") pod \"3d7ff16e-2ecf-44de-9b10-5c60981d674c\" (UID: \"3d7ff16e-2ecf-44de-9b10-5c60981d674c\") " Dec 01 09:03:49 crc kubenswrapper[4873]: I1201 09:03:49.394181 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3d7ff16e-2ecf-44de-9b10-5c60981d674c-ovsdbserver-nb\") pod \"3d7ff16e-2ecf-44de-9b10-5c60981d674c\" (UID: \"3d7ff16e-2ecf-44de-9b10-5c60981d674c\") " Dec 01 09:03:49 crc kubenswrapper[4873]: I1201 09:03:49.394250 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d7ff16e-2ecf-44de-9b10-5c60981d674c-config\") pod \"3d7ff16e-2ecf-44de-9b10-5c60981d674c\" (UID: \"3d7ff16e-2ecf-44de-9b10-5c60981d674c\") " Dec 01 09:03:49 crc kubenswrapper[4873]: I1201 09:03:49.394287 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q46ff\" (UniqueName: \"kubernetes.io/projected/3d7ff16e-2ecf-44de-9b10-5c60981d674c-kube-api-access-q46ff\") pod \"3d7ff16e-2ecf-44de-9b10-5c60981d674c\" (UID: \"3d7ff16e-2ecf-44de-9b10-5c60981d674c\") " Dec 01 09:03:49 crc kubenswrapper[4873]: I1201 09:03:49.394312 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3d7ff16e-2ecf-44de-9b10-5c60981d674c-dns-svc\") pod \"3d7ff16e-2ecf-44de-9b10-5c60981d674c\" (UID: \"3d7ff16e-2ecf-44de-9b10-5c60981d674c\") " Dec 01 09:03:49 crc kubenswrapper[4873]: I1201 09:03:49.401830 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d7ff16e-2ecf-44de-9b10-5c60981d674c-kube-api-access-q46ff" (OuterVolumeSpecName: "kube-api-access-q46ff") pod "3d7ff16e-2ecf-44de-9b10-5c60981d674c" (UID: "3d7ff16e-2ecf-44de-9b10-5c60981d674c"). InnerVolumeSpecName "kube-api-access-q46ff". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:03:49 crc kubenswrapper[4873]: I1201 09:03:49.434260 4873 generic.go:334] "Generic (PLEG): container finished" podID="3d7ff16e-2ecf-44de-9b10-5c60981d674c" containerID="da46494a66f06ddde14ed0184bcadb65bcb1085b8fa27f8069d81792116386b1" exitCode=0 Dec 01 09:03:49 crc kubenswrapper[4873]: I1201 09:03:49.434600 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6447ccbd8f-rgfqr" event={"ID":"3d7ff16e-2ecf-44de-9b10-5c60981d674c","Type":"ContainerDied","Data":"da46494a66f06ddde14ed0184bcadb65bcb1085b8fa27f8069d81792116386b1"} Dec 01 09:03:49 crc kubenswrapper[4873]: I1201 09:03:49.434691 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6447ccbd8f-rgfqr" event={"ID":"3d7ff16e-2ecf-44de-9b10-5c60981d674c","Type":"ContainerDied","Data":"c8aef672ba09680bc8ac2afcfd4102a5b972bc9bfdf8e83a1a07c5ab02f994a4"} Dec 01 09:03:49 crc kubenswrapper[4873]: I1201 09:03:49.434771 4873 scope.go:117] "RemoveContainer" containerID="da46494a66f06ddde14ed0184bcadb65bcb1085b8fa27f8069d81792116386b1" Dec 01 09:03:49 crc kubenswrapper[4873]: I1201 09:03:49.434988 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6447ccbd8f-rgfqr" Dec 01 09:03:49 crc kubenswrapper[4873]: I1201 09:03:49.465500 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d7ff16e-2ecf-44de-9b10-5c60981d674c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3d7ff16e-2ecf-44de-9b10-5c60981d674c" (UID: "3d7ff16e-2ecf-44de-9b10-5c60981d674c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:03:49 crc kubenswrapper[4873]: I1201 09:03:49.468727 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d7ff16e-2ecf-44de-9b10-5c60981d674c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3d7ff16e-2ecf-44de-9b10-5c60981d674c" (UID: "3d7ff16e-2ecf-44de-9b10-5c60981d674c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:03:49 crc kubenswrapper[4873]: I1201 09:03:49.472339 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d7ff16e-2ecf-44de-9b10-5c60981d674c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3d7ff16e-2ecf-44de-9b10-5c60981d674c" (UID: "3d7ff16e-2ecf-44de-9b10-5c60981d674c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:03:49 crc kubenswrapper[4873]: I1201 09:03:49.474174 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d7ff16e-2ecf-44de-9b10-5c60981d674c-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "3d7ff16e-2ecf-44de-9b10-5c60981d674c" (UID: "3d7ff16e-2ecf-44de-9b10-5c60981d674c"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:03:49 crc kubenswrapper[4873]: I1201 09:03:49.485220 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d7ff16e-2ecf-44de-9b10-5c60981d674c-config" (OuterVolumeSpecName: "config") pod "3d7ff16e-2ecf-44de-9b10-5c60981d674c" (UID: "3d7ff16e-2ecf-44de-9b10-5c60981d674c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:03:49 crc kubenswrapper[4873]: I1201 09:03:49.497532 4873 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3d7ff16e-2ecf-44de-9b10-5c60981d674c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:49 crc kubenswrapper[4873]: I1201 09:03:49.497586 4873 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/3d7ff16e-2ecf-44de-9b10-5c60981d674c-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:49 crc kubenswrapper[4873]: I1201 09:03:49.497601 4873 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3d7ff16e-2ecf-44de-9b10-5c60981d674c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:49 crc kubenswrapper[4873]: I1201 09:03:49.497615 4873 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d7ff16e-2ecf-44de-9b10-5c60981d674c-config\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:49 crc kubenswrapper[4873]: I1201 09:03:49.497633 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q46ff\" (UniqueName: \"kubernetes.io/projected/3d7ff16e-2ecf-44de-9b10-5c60981d674c-kube-api-access-q46ff\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:49 crc kubenswrapper[4873]: I1201 09:03:49.497648 4873 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3d7ff16e-2ecf-44de-9b10-5c60981d674c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:49 crc kubenswrapper[4873]: I1201 09:03:49.557840 4873 scope.go:117] "RemoveContainer" containerID="ba0b6bc5473f347d003e0106e0fbe6b2edc21402e8b91ad650e4348efa779653" Dec 01 09:03:49 crc kubenswrapper[4873]: I1201 09:03:49.583711 4873 scope.go:117] "RemoveContainer" containerID="da46494a66f06ddde14ed0184bcadb65bcb1085b8fa27f8069d81792116386b1" Dec 01 09:03:49 crc kubenswrapper[4873]: E1201 09:03:49.584299 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da46494a66f06ddde14ed0184bcadb65bcb1085b8fa27f8069d81792116386b1\": container with ID starting with da46494a66f06ddde14ed0184bcadb65bcb1085b8fa27f8069d81792116386b1 not found: ID does not exist" containerID="da46494a66f06ddde14ed0184bcadb65bcb1085b8fa27f8069d81792116386b1" Dec 01 09:03:49 crc kubenswrapper[4873]: I1201 09:03:49.584362 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da46494a66f06ddde14ed0184bcadb65bcb1085b8fa27f8069d81792116386b1"} err="failed to get container status \"da46494a66f06ddde14ed0184bcadb65bcb1085b8fa27f8069d81792116386b1\": rpc error: code = NotFound desc = could not find container \"da46494a66f06ddde14ed0184bcadb65bcb1085b8fa27f8069d81792116386b1\": container with ID starting with da46494a66f06ddde14ed0184bcadb65bcb1085b8fa27f8069d81792116386b1 not found: ID does not exist" Dec 01 09:03:49 crc kubenswrapper[4873]: I1201 09:03:49.584392 4873 scope.go:117] "RemoveContainer" containerID="ba0b6bc5473f347d003e0106e0fbe6b2edc21402e8b91ad650e4348efa779653" Dec 01 09:03:49 crc kubenswrapper[4873]: E1201 09:03:49.584705 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba0b6bc5473f347d003e0106e0fbe6b2edc21402e8b91ad650e4348efa779653\": container with ID starting with ba0b6bc5473f347d003e0106e0fbe6b2edc21402e8b91ad650e4348efa779653 not found: ID does not exist" containerID="ba0b6bc5473f347d003e0106e0fbe6b2edc21402e8b91ad650e4348efa779653" Dec 01 09:03:49 crc kubenswrapper[4873]: I1201 09:03:49.584745 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba0b6bc5473f347d003e0106e0fbe6b2edc21402e8b91ad650e4348efa779653"} err="failed to get container status \"ba0b6bc5473f347d003e0106e0fbe6b2edc21402e8b91ad650e4348efa779653\": rpc error: code = NotFound desc = could not find container \"ba0b6bc5473f347d003e0106e0fbe6b2edc21402e8b91ad650e4348efa779653\": container with ID starting with ba0b6bc5473f347d003e0106e0fbe6b2edc21402e8b91ad650e4348efa779653 not found: ID does not exist" Dec 01 09:03:49 crc kubenswrapper[4873]: I1201 09:03:49.770317 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6447ccbd8f-rgfqr"] Dec 01 09:03:49 crc kubenswrapper[4873]: I1201 09:03:49.780138 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6447ccbd8f-rgfqr"] Dec 01 09:03:50 crc kubenswrapper[4873]: I1201 09:03:50.462963 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d7ff16e-2ecf-44de-9b10-5c60981d674c" path="/var/lib/kubelet/pods/3d7ff16e-2ecf-44de-9b10-5c60981d674c/volumes" Dec 01 09:03:58 crc kubenswrapper[4873]: I1201 09:03:58.664688 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wfjd2"] Dec 01 09:03:58 crc kubenswrapper[4873]: E1201 09:03:58.666477 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d7ff16e-2ecf-44de-9b10-5c60981d674c" containerName="init" Dec 01 09:03:58 crc kubenswrapper[4873]: I1201 09:03:58.666514 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d7ff16e-2ecf-44de-9b10-5c60981d674c" containerName="init" Dec 01 09:03:58 crc kubenswrapper[4873]: E1201 09:03:58.666551 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d7ff16e-2ecf-44de-9b10-5c60981d674c" containerName="dnsmasq-dns" Dec 01 09:03:58 crc kubenswrapper[4873]: I1201 09:03:58.666569 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d7ff16e-2ecf-44de-9b10-5c60981d674c" containerName="dnsmasq-dns" Dec 01 09:03:58 crc kubenswrapper[4873]: E1201 09:03:58.666598 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e825def4-1364-4a5b-8896-b8b633a8142c" containerName="dnsmasq-dns" Dec 01 09:03:58 crc kubenswrapper[4873]: I1201 09:03:58.666618 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="e825def4-1364-4a5b-8896-b8b633a8142c" containerName="dnsmasq-dns" Dec 01 09:03:58 crc kubenswrapper[4873]: E1201 09:03:58.666687 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e825def4-1364-4a5b-8896-b8b633a8142c" containerName="init" Dec 01 09:03:58 crc kubenswrapper[4873]: I1201 09:03:58.666703 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="e825def4-1364-4a5b-8896-b8b633a8142c" containerName="init" Dec 01 09:03:58 crc kubenswrapper[4873]: I1201 09:03:58.667173 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="e825def4-1364-4a5b-8896-b8b633a8142c" containerName="dnsmasq-dns" Dec 01 09:03:58 crc kubenswrapper[4873]: I1201 09:03:58.667200 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d7ff16e-2ecf-44de-9b10-5c60981d674c" containerName="dnsmasq-dns" Dec 01 09:03:58 crc kubenswrapper[4873]: I1201 09:03:58.668634 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wfjd2" Dec 01 09:03:58 crc kubenswrapper[4873]: I1201 09:03:58.692218 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wfjd2"] Dec 01 09:03:58 crc kubenswrapper[4873]: I1201 09:03:58.704572 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 09:03:58 crc kubenswrapper[4873]: I1201 09:03:58.705908 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9ghqw" Dec 01 09:03:58 crc kubenswrapper[4873]: I1201 09:03:58.706008 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 09:03:58 crc kubenswrapper[4873]: I1201 09:03:58.706003 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 09:03:58 crc kubenswrapper[4873]: I1201 09:03:58.820772 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smt6j\" (UniqueName: \"kubernetes.io/projected/ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd-kube-api-access-smt6j\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wfjd2\" (UID: \"ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wfjd2" Dec 01 09:03:58 crc kubenswrapper[4873]: I1201 09:03:58.820855 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wfjd2\" (UID: \"ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wfjd2" Dec 01 09:03:58 crc kubenswrapper[4873]: I1201 09:03:58.821090 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wfjd2\" (UID: \"ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wfjd2" Dec 01 09:03:58 crc kubenswrapper[4873]: I1201 09:03:58.821178 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wfjd2\" (UID: \"ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wfjd2" Dec 01 09:03:58 crc kubenswrapper[4873]: I1201 09:03:58.923108 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smt6j\" (UniqueName: \"kubernetes.io/projected/ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd-kube-api-access-smt6j\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wfjd2\" (UID: \"ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wfjd2" Dec 01 09:03:58 crc kubenswrapper[4873]: I1201 09:03:58.923204 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wfjd2\" (UID: \"ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wfjd2" Dec 01 09:03:58 crc kubenswrapper[4873]: I1201 09:03:58.923286 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wfjd2\" (UID: \"ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wfjd2" Dec 01 09:03:58 crc kubenswrapper[4873]: I1201 09:03:58.923335 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wfjd2\" (UID: \"ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wfjd2" Dec 01 09:03:58 crc kubenswrapper[4873]: I1201 09:03:58.933735 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wfjd2\" (UID: \"ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wfjd2" Dec 01 09:03:58 crc kubenswrapper[4873]: I1201 09:03:58.933941 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wfjd2\" (UID: \"ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wfjd2" Dec 01 09:03:58 crc kubenswrapper[4873]: I1201 09:03:58.934223 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wfjd2\" (UID: \"ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wfjd2" Dec 01 09:03:58 crc kubenswrapper[4873]: I1201 09:03:58.945191 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smt6j\" (UniqueName: \"kubernetes.io/projected/ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd-kube-api-access-smt6j\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wfjd2\" (UID: \"ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wfjd2" Dec 01 09:03:59 crc kubenswrapper[4873]: I1201 09:03:59.021637 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wfjd2" Dec 01 09:03:59 crc kubenswrapper[4873]: I1201 09:03:59.419393 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wfjd2"] Dec 01 09:03:59 crc kubenswrapper[4873]: W1201 09:03:59.429345 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podba3d9d8a_e00b_467f_b718_d5fcfeea1fdd.slice/crio-0361c2d45555a73e27df36f8d51a3dae331afb938a8683ad7af47d99734904b9 WatchSource:0}: Error finding container 0361c2d45555a73e27df36f8d51a3dae331afb938a8683ad7af47d99734904b9: Status 404 returned error can't find the container with id 0361c2d45555a73e27df36f8d51a3dae331afb938a8683ad7af47d99734904b9 Dec 01 09:03:59 crc kubenswrapper[4873]: I1201 09:03:59.563669 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wfjd2" event={"ID":"ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd","Type":"ContainerStarted","Data":"0361c2d45555a73e27df36f8d51a3dae331afb938a8683ad7af47d99734904b9"} Dec 01 09:04:00 crc kubenswrapper[4873]: I1201 09:04:00.575460 4873 generic.go:334] "Generic (PLEG): container finished" podID="84aac4b8-9ed2-40d2-81f1-eba3af6e46b0" containerID="f6ecd62b7b64755500accd3a857b33968f234e31b8cb9d782b3e3626cf2fd48a" exitCode=0 Dec 01 09:04:00 crc kubenswrapper[4873]: I1201 09:04:00.575517 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"84aac4b8-9ed2-40d2-81f1-eba3af6e46b0","Type":"ContainerDied","Data":"f6ecd62b7b64755500accd3a857b33968f234e31b8cb9d782b3e3626cf2fd48a"} Dec 01 09:04:01 crc kubenswrapper[4873]: I1201 09:04:01.587304 4873 generic.go:334] "Generic (PLEG): container finished" podID="3baa8150-922b-4de7-ae44-24b9b5584212" containerID="d6285fe2b3b994bd13a0add987955acea8a299727a6960e9c0a06224b4764e50" exitCode=0 Dec 01 09:04:01 crc kubenswrapper[4873]: I1201 09:04:01.587369 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3baa8150-922b-4de7-ae44-24b9b5584212","Type":"ContainerDied","Data":"d6285fe2b3b994bd13a0add987955acea8a299727a6960e9c0a06224b4764e50"} Dec 01 09:04:01 crc kubenswrapper[4873]: I1201 09:04:01.589799 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"84aac4b8-9ed2-40d2-81f1-eba3af6e46b0","Type":"ContainerStarted","Data":"172424c9115d0112ca874a67c4b8a1bb2323cd326ac2139af16cfcfa0ac55876"} Dec 01 09:04:01 crc kubenswrapper[4873]: I1201 09:04:01.590384 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 01 09:04:01 crc kubenswrapper[4873]: I1201 09:04:01.645578 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.645545231 podStartE2EDuration="37.645545231s" podCreationTimestamp="2025-12-01 09:03:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:04:01.64021109 +0000 UTC m=+1417.542319629" watchObservedRunningTime="2025-12-01 09:04:01.645545231 +0000 UTC m=+1417.547653780" Dec 01 09:04:02 crc kubenswrapper[4873]: I1201 09:04:02.601840 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3baa8150-922b-4de7-ae44-24b9b5584212","Type":"ContainerStarted","Data":"4e0dbf880b26206f42026a20fdf6b437899b6e017f7f9549339c55af82df3b4b"} Dec 01 09:04:02 crc kubenswrapper[4873]: I1201 09:04:02.602257 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 01 09:04:02 crc kubenswrapper[4873]: I1201 09:04:02.634836 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.634798814 podStartE2EDuration="37.634798814s" podCreationTimestamp="2025-12-01 09:03:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:04:02.625541088 +0000 UTC m=+1418.527649637" watchObservedRunningTime="2025-12-01 09:04:02.634798814 +0000 UTC m=+1418.536907363" Dec 01 09:04:10 crc kubenswrapper[4873]: I1201 09:04:10.698545 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wfjd2" event={"ID":"ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd","Type":"ContainerStarted","Data":"dfac8b2123466a36a436429e744cce54ec673ca0e1a5b3c80ea5a8ace985a68b"} Dec 01 09:04:10 crc kubenswrapper[4873]: I1201 09:04:10.718180 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wfjd2" podStartSLOduration=2.194065614 podStartE2EDuration="12.71815785s" podCreationTimestamp="2025-12-01 09:03:58 +0000 UTC" firstStartedPulling="2025-12-01 09:03:59.433077281 +0000 UTC m=+1415.335185830" lastFinishedPulling="2025-12-01 09:04:09.957169527 +0000 UTC m=+1425.859278066" observedRunningTime="2025-12-01 09:04:10.71733969 +0000 UTC m=+1426.619448259" watchObservedRunningTime="2025-12-01 09:04:10.71815785 +0000 UTC m=+1426.620266389" Dec 01 09:04:14 crc kubenswrapper[4873]: I1201 09:04:14.444217 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 01 09:04:15 crc kubenswrapper[4873]: I1201 09:04:15.552288 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 01 09:04:21 crc kubenswrapper[4873]: I1201 09:04:21.843666 4873 generic.go:334] "Generic (PLEG): container finished" podID="ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd" containerID="dfac8b2123466a36a436429e744cce54ec673ca0e1a5b3c80ea5a8ace985a68b" exitCode=0 Dec 01 09:04:21 crc kubenswrapper[4873]: I1201 09:04:21.843924 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wfjd2" event={"ID":"ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd","Type":"ContainerDied","Data":"dfac8b2123466a36a436429e744cce54ec673ca0e1a5b3c80ea5a8ace985a68b"} Dec 01 09:04:23 crc kubenswrapper[4873]: I1201 09:04:23.385163 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wfjd2" Dec 01 09:04:23 crc kubenswrapper[4873]: I1201 09:04:23.590334 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd-ssh-key\") pod \"ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd\" (UID: \"ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd\") " Dec 01 09:04:23 crc kubenswrapper[4873]: I1201 09:04:23.590450 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd-inventory\") pod \"ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd\" (UID: \"ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd\") " Dec 01 09:04:23 crc kubenswrapper[4873]: I1201 09:04:23.590481 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd-repo-setup-combined-ca-bundle\") pod \"ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd\" (UID: \"ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd\") " Dec 01 09:04:23 crc kubenswrapper[4873]: I1201 09:04:23.591595 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-smt6j\" (UniqueName: \"kubernetes.io/projected/ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd-kube-api-access-smt6j\") pod \"ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd\" (UID: \"ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd\") " Dec 01 09:04:23 crc kubenswrapper[4873]: I1201 09:04:23.598147 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd-kube-api-access-smt6j" (OuterVolumeSpecName: "kube-api-access-smt6j") pod "ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd" (UID: "ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd"). InnerVolumeSpecName "kube-api-access-smt6j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:04:23 crc kubenswrapper[4873]: I1201 09:04:23.598225 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd" (UID: "ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:04:23 crc kubenswrapper[4873]: I1201 09:04:23.619947 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd" (UID: "ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:04:23 crc kubenswrapper[4873]: I1201 09:04:23.628618 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd-inventory" (OuterVolumeSpecName: "inventory") pod "ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd" (UID: "ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:04:23 crc kubenswrapper[4873]: I1201 09:04:23.694621 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-smt6j\" (UniqueName: \"kubernetes.io/projected/ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd-kube-api-access-smt6j\") on node \"crc\" DevicePath \"\"" Dec 01 09:04:23 crc kubenswrapper[4873]: I1201 09:04:23.694903 4873 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 09:04:23 crc kubenswrapper[4873]: I1201 09:04:23.694972 4873 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 09:04:23 crc kubenswrapper[4873]: I1201 09:04:23.695088 4873 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:04:23 crc kubenswrapper[4873]: I1201 09:04:23.871191 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wfjd2" event={"ID":"ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd","Type":"ContainerDied","Data":"0361c2d45555a73e27df36f8d51a3dae331afb938a8683ad7af47d99734904b9"} Dec 01 09:04:23 crc kubenswrapper[4873]: I1201 09:04:23.871234 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wfjd2" Dec 01 09:04:23 crc kubenswrapper[4873]: I1201 09:04:23.871252 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0361c2d45555a73e27df36f8d51a3dae331afb938a8683ad7af47d99734904b9" Dec 01 09:04:23 crc kubenswrapper[4873]: I1201 09:04:23.952941 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-f9cnt"] Dec 01 09:04:23 crc kubenswrapper[4873]: E1201 09:04:23.953452 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 01 09:04:23 crc kubenswrapper[4873]: I1201 09:04:23.953474 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 01 09:04:23 crc kubenswrapper[4873]: I1201 09:04:23.953670 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 01 09:04:23 crc kubenswrapper[4873]: I1201 09:04:23.954398 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-f9cnt" Dec 01 09:04:23 crc kubenswrapper[4873]: I1201 09:04:23.956611 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 09:04:23 crc kubenswrapper[4873]: I1201 09:04:23.956819 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9ghqw" Dec 01 09:04:23 crc kubenswrapper[4873]: I1201 09:04:23.956911 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 09:04:23 crc kubenswrapper[4873]: I1201 09:04:23.956961 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 09:04:23 crc kubenswrapper[4873]: I1201 09:04:23.962394 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-f9cnt"] Dec 01 09:04:24 crc kubenswrapper[4873]: I1201 09:04:24.000417 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/489b7faf-eb94-459f-b279-7b5cedf53db9-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-f9cnt\" (UID: \"489b7faf-eb94-459f-b279-7b5cedf53db9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-f9cnt" Dec 01 09:04:24 crc kubenswrapper[4873]: I1201 09:04:24.000472 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/489b7faf-eb94-459f-b279-7b5cedf53db9-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-f9cnt\" (UID: \"489b7faf-eb94-459f-b279-7b5cedf53db9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-f9cnt" Dec 01 09:04:24 crc kubenswrapper[4873]: I1201 09:04:24.000612 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hh22d\" (UniqueName: \"kubernetes.io/projected/489b7faf-eb94-459f-b279-7b5cedf53db9-kube-api-access-hh22d\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-f9cnt\" (UID: \"489b7faf-eb94-459f-b279-7b5cedf53db9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-f9cnt" Dec 01 09:04:24 crc kubenswrapper[4873]: I1201 09:04:24.000665 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/489b7faf-eb94-459f-b279-7b5cedf53db9-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-f9cnt\" (UID: \"489b7faf-eb94-459f-b279-7b5cedf53db9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-f9cnt" Dec 01 09:04:24 crc kubenswrapper[4873]: I1201 09:04:24.102445 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/489b7faf-eb94-459f-b279-7b5cedf53db9-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-f9cnt\" (UID: \"489b7faf-eb94-459f-b279-7b5cedf53db9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-f9cnt" Dec 01 09:04:24 crc kubenswrapper[4873]: I1201 09:04:24.102499 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/489b7faf-eb94-459f-b279-7b5cedf53db9-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-f9cnt\" (UID: \"489b7faf-eb94-459f-b279-7b5cedf53db9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-f9cnt" Dec 01 09:04:24 crc kubenswrapper[4873]: I1201 09:04:24.102564 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hh22d\" (UniqueName: \"kubernetes.io/projected/489b7faf-eb94-459f-b279-7b5cedf53db9-kube-api-access-hh22d\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-f9cnt\" (UID: \"489b7faf-eb94-459f-b279-7b5cedf53db9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-f9cnt" Dec 01 09:04:24 crc kubenswrapper[4873]: I1201 09:04:24.102620 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/489b7faf-eb94-459f-b279-7b5cedf53db9-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-f9cnt\" (UID: \"489b7faf-eb94-459f-b279-7b5cedf53db9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-f9cnt" Dec 01 09:04:24 crc kubenswrapper[4873]: I1201 09:04:24.110374 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/489b7faf-eb94-459f-b279-7b5cedf53db9-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-f9cnt\" (UID: \"489b7faf-eb94-459f-b279-7b5cedf53db9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-f9cnt" Dec 01 09:04:24 crc kubenswrapper[4873]: I1201 09:04:24.110581 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/489b7faf-eb94-459f-b279-7b5cedf53db9-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-f9cnt\" (UID: \"489b7faf-eb94-459f-b279-7b5cedf53db9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-f9cnt" Dec 01 09:04:24 crc kubenswrapper[4873]: I1201 09:04:24.115165 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/489b7faf-eb94-459f-b279-7b5cedf53db9-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-f9cnt\" (UID: \"489b7faf-eb94-459f-b279-7b5cedf53db9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-f9cnt" Dec 01 09:04:24 crc kubenswrapper[4873]: I1201 09:04:24.132198 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hh22d\" (UniqueName: \"kubernetes.io/projected/489b7faf-eb94-459f-b279-7b5cedf53db9-kube-api-access-hh22d\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-f9cnt\" (UID: \"489b7faf-eb94-459f-b279-7b5cedf53db9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-f9cnt" Dec 01 09:04:24 crc kubenswrapper[4873]: I1201 09:04:24.284888 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-f9cnt" Dec 01 09:04:24 crc kubenswrapper[4873]: I1201 09:04:24.874982 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-f9cnt"] Dec 01 09:04:24 crc kubenswrapper[4873]: W1201 09:04:24.878043 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod489b7faf_eb94_459f_b279_7b5cedf53db9.slice/crio-adca14a3f7ed588e4e72c1327a3b999ba237a32e2b06efa67686a0d065cecf8e WatchSource:0}: Error finding container adca14a3f7ed588e4e72c1327a3b999ba237a32e2b06efa67686a0d065cecf8e: Status 404 returned error can't find the container with id adca14a3f7ed588e4e72c1327a3b999ba237a32e2b06efa67686a0d065cecf8e Dec 01 09:04:25 crc kubenswrapper[4873]: I1201 09:04:25.888076 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-f9cnt" event={"ID":"489b7faf-eb94-459f-b279-7b5cedf53db9","Type":"ContainerStarted","Data":"adca14a3f7ed588e4e72c1327a3b999ba237a32e2b06efa67686a0d065cecf8e"} Dec 01 09:04:26 crc kubenswrapper[4873]: I1201 09:04:26.383829 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 09:04:26 crc kubenswrapper[4873]: I1201 09:04:26.898430 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-f9cnt" event={"ID":"489b7faf-eb94-459f-b279-7b5cedf53db9","Type":"ContainerStarted","Data":"0b17315b3b408f1a71b69372e87fd41c476a1d6148d2474063bce7ce3f666609"} Dec 01 09:04:26 crc kubenswrapper[4873]: I1201 09:04:26.924510 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-f9cnt" podStartSLOduration=2.42525903 podStartE2EDuration="3.924465958s" podCreationTimestamp="2025-12-01 09:04:23 +0000 UTC" firstStartedPulling="2025-12-01 09:04:24.881224264 +0000 UTC m=+1440.783332803" lastFinishedPulling="2025-12-01 09:04:26.380431192 +0000 UTC m=+1442.282539731" observedRunningTime="2025-12-01 09:04:26.915247755 +0000 UTC m=+1442.817356294" watchObservedRunningTime="2025-12-01 09:04:26.924465958 +0000 UTC m=+1442.826574537" Dec 01 09:05:01 crc kubenswrapper[4873]: I1201 09:05:01.058946 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:05:01 crc kubenswrapper[4873]: I1201 09:05:01.060113 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:05:31 crc kubenswrapper[4873]: I1201 09:05:31.059682 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:05:31 crc kubenswrapper[4873]: I1201 09:05:31.060854 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:05:55 crc kubenswrapper[4873]: I1201 09:05:55.365992 4873 scope.go:117] "RemoveContainer" containerID="8290cdaf32cb1fefd00caf645207c20e538df3ee40573819f2a2ee25f0f677e5" Dec 01 09:05:56 crc kubenswrapper[4873]: I1201 09:05:56.032631 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-46xcn"] Dec 01 09:05:56 crc kubenswrapper[4873]: I1201 09:05:56.036576 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-46xcn" Dec 01 09:05:56 crc kubenswrapper[4873]: I1201 09:05:56.052969 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-46xcn"] Dec 01 09:05:56 crc kubenswrapper[4873]: I1201 09:05:56.125174 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9aa84f5-b53c-4a95-8204-03128df83d58-utilities\") pod \"redhat-marketplace-46xcn\" (UID: \"f9aa84f5-b53c-4a95-8204-03128df83d58\") " pod="openshift-marketplace/redhat-marketplace-46xcn" Dec 01 09:05:56 crc kubenswrapper[4873]: I1201 09:05:56.125236 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9aa84f5-b53c-4a95-8204-03128df83d58-catalog-content\") pod \"redhat-marketplace-46xcn\" (UID: \"f9aa84f5-b53c-4a95-8204-03128df83d58\") " pod="openshift-marketplace/redhat-marketplace-46xcn" Dec 01 09:05:56 crc kubenswrapper[4873]: I1201 09:05:56.125294 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5xjn\" (UniqueName: \"kubernetes.io/projected/f9aa84f5-b53c-4a95-8204-03128df83d58-kube-api-access-w5xjn\") pod \"redhat-marketplace-46xcn\" (UID: \"f9aa84f5-b53c-4a95-8204-03128df83d58\") " pod="openshift-marketplace/redhat-marketplace-46xcn" Dec 01 09:05:56 crc kubenswrapper[4873]: I1201 09:05:56.228106 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9aa84f5-b53c-4a95-8204-03128df83d58-utilities\") pod \"redhat-marketplace-46xcn\" (UID: \"f9aa84f5-b53c-4a95-8204-03128df83d58\") " pod="openshift-marketplace/redhat-marketplace-46xcn" Dec 01 09:05:56 crc kubenswrapper[4873]: I1201 09:05:56.228169 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9aa84f5-b53c-4a95-8204-03128df83d58-catalog-content\") pod \"redhat-marketplace-46xcn\" (UID: \"f9aa84f5-b53c-4a95-8204-03128df83d58\") " pod="openshift-marketplace/redhat-marketplace-46xcn" Dec 01 09:05:56 crc kubenswrapper[4873]: I1201 09:05:56.228194 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5xjn\" (UniqueName: \"kubernetes.io/projected/f9aa84f5-b53c-4a95-8204-03128df83d58-kube-api-access-w5xjn\") pod \"redhat-marketplace-46xcn\" (UID: \"f9aa84f5-b53c-4a95-8204-03128df83d58\") " pod="openshift-marketplace/redhat-marketplace-46xcn" Dec 01 09:05:56 crc kubenswrapper[4873]: I1201 09:05:56.229199 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9aa84f5-b53c-4a95-8204-03128df83d58-catalog-content\") pod \"redhat-marketplace-46xcn\" (UID: \"f9aa84f5-b53c-4a95-8204-03128df83d58\") " pod="openshift-marketplace/redhat-marketplace-46xcn" Dec 01 09:05:56 crc kubenswrapper[4873]: I1201 09:05:56.229273 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9aa84f5-b53c-4a95-8204-03128df83d58-utilities\") pod \"redhat-marketplace-46xcn\" (UID: \"f9aa84f5-b53c-4a95-8204-03128df83d58\") " pod="openshift-marketplace/redhat-marketplace-46xcn" Dec 01 09:05:56 crc kubenswrapper[4873]: I1201 09:05:56.262540 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5xjn\" (UniqueName: \"kubernetes.io/projected/f9aa84f5-b53c-4a95-8204-03128df83d58-kube-api-access-w5xjn\") pod \"redhat-marketplace-46xcn\" (UID: \"f9aa84f5-b53c-4a95-8204-03128df83d58\") " pod="openshift-marketplace/redhat-marketplace-46xcn" Dec 01 09:05:56 crc kubenswrapper[4873]: I1201 09:05:56.365616 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-46xcn" Dec 01 09:05:56 crc kubenswrapper[4873]: I1201 09:05:56.902457 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-46xcn"] Dec 01 09:05:56 crc kubenswrapper[4873]: W1201 09:05:56.911316 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf9aa84f5_b53c_4a95_8204_03128df83d58.slice/crio-d390b43835fabc37d992ed00f2d905bfdfd5cf7dc057132605641673305607a5 WatchSource:0}: Error finding container d390b43835fabc37d992ed00f2d905bfdfd5cf7dc057132605641673305607a5: Status 404 returned error can't find the container with id d390b43835fabc37d992ed00f2d905bfdfd5cf7dc057132605641673305607a5 Dec 01 09:05:57 crc kubenswrapper[4873]: I1201 09:05:57.205519 4873 generic.go:334] "Generic (PLEG): container finished" podID="f9aa84f5-b53c-4a95-8204-03128df83d58" containerID="0a10f69ef3b1214f2868cb957b4c6f9b1393e665089901b3ea3120c4b3e3ed74" exitCode=0 Dec 01 09:05:57 crc kubenswrapper[4873]: I1201 09:05:57.205634 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-46xcn" event={"ID":"f9aa84f5-b53c-4a95-8204-03128df83d58","Type":"ContainerDied","Data":"0a10f69ef3b1214f2868cb957b4c6f9b1393e665089901b3ea3120c4b3e3ed74"} Dec 01 09:05:57 crc kubenswrapper[4873]: I1201 09:05:57.205958 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-46xcn" event={"ID":"f9aa84f5-b53c-4a95-8204-03128df83d58","Type":"ContainerStarted","Data":"d390b43835fabc37d992ed00f2d905bfdfd5cf7dc057132605641673305607a5"} Dec 01 09:05:58 crc kubenswrapper[4873]: I1201 09:05:58.219777 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-46xcn" event={"ID":"f9aa84f5-b53c-4a95-8204-03128df83d58","Type":"ContainerStarted","Data":"a3d2a9684a9f004b15da114e3d120eaa142ea65d58afc36b7052310b36ebe652"} Dec 01 09:05:59 crc kubenswrapper[4873]: I1201 09:05:59.231195 4873 generic.go:334] "Generic (PLEG): container finished" podID="f9aa84f5-b53c-4a95-8204-03128df83d58" containerID="a3d2a9684a9f004b15da114e3d120eaa142ea65d58afc36b7052310b36ebe652" exitCode=0 Dec 01 09:05:59 crc kubenswrapper[4873]: I1201 09:05:59.231322 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-46xcn" event={"ID":"f9aa84f5-b53c-4a95-8204-03128df83d58","Type":"ContainerDied","Data":"a3d2a9684a9f004b15da114e3d120eaa142ea65d58afc36b7052310b36ebe652"} Dec 01 09:05:59 crc kubenswrapper[4873]: I1201 09:05:59.231592 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-46xcn" event={"ID":"f9aa84f5-b53c-4a95-8204-03128df83d58","Type":"ContainerStarted","Data":"b3723557d214bc04f9dc734d9fb0626d6b67632eb283c7c3b65da4bb8f045eb9"} Dec 01 09:05:59 crc kubenswrapper[4873]: I1201 09:05:59.262447 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-46xcn" podStartSLOduration=2.699080398 podStartE2EDuration="4.262420133s" podCreationTimestamp="2025-12-01 09:05:55 +0000 UTC" firstStartedPulling="2025-12-01 09:05:57.207117689 +0000 UTC m=+1533.109226228" lastFinishedPulling="2025-12-01 09:05:58.770457414 +0000 UTC m=+1534.672565963" observedRunningTime="2025-12-01 09:05:59.255061235 +0000 UTC m=+1535.157169784" watchObservedRunningTime="2025-12-01 09:05:59.262420133 +0000 UTC m=+1535.164528672" Dec 01 09:06:01 crc kubenswrapper[4873]: I1201 09:06:01.059385 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:06:01 crc kubenswrapper[4873]: I1201 09:06:01.059877 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:06:01 crc kubenswrapper[4873]: I1201 09:06:01.059935 4873 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" Dec 01 09:06:01 crc kubenswrapper[4873]: I1201 09:06:01.061133 4873 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dd1fd5c773938eaca06d208e8759b6376ff65c7cad3dc995d4b7c31e75f098dc"} pod="openshift-machine-config-operator/machine-config-daemon-scwpp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 09:06:01 crc kubenswrapper[4873]: I1201 09:06:01.061219 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" containerID="cri-o://dd1fd5c773938eaca06d208e8759b6376ff65c7cad3dc995d4b7c31e75f098dc" gracePeriod=600 Dec 01 09:06:01 crc kubenswrapper[4873]: E1201 09:06:01.189725 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:06:01 crc kubenswrapper[4873]: I1201 09:06:01.265768 4873 generic.go:334] "Generic (PLEG): container finished" podID="fef7b114-0e07-402d-a37b-315c36011f4b" containerID="dd1fd5c773938eaca06d208e8759b6376ff65c7cad3dc995d4b7c31e75f098dc" exitCode=0 Dec 01 09:06:01 crc kubenswrapper[4873]: I1201 09:06:01.265854 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" event={"ID":"fef7b114-0e07-402d-a37b-315c36011f4b","Type":"ContainerDied","Data":"dd1fd5c773938eaca06d208e8759b6376ff65c7cad3dc995d4b7c31e75f098dc"} Dec 01 09:06:01 crc kubenswrapper[4873]: I1201 09:06:01.265964 4873 scope.go:117] "RemoveContainer" containerID="90066b4e910f0ee21a175275171e35716a5459a2c723385e352a837980db5011" Dec 01 09:06:01 crc kubenswrapper[4873]: I1201 09:06:01.266797 4873 scope.go:117] "RemoveContainer" containerID="dd1fd5c773938eaca06d208e8759b6376ff65c7cad3dc995d4b7c31e75f098dc" Dec 01 09:06:01 crc kubenswrapper[4873]: E1201 09:06:01.267143 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:06:06 crc kubenswrapper[4873]: I1201 09:06:06.366345 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-46xcn" Dec 01 09:06:06 crc kubenswrapper[4873]: I1201 09:06:06.367235 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-46xcn" Dec 01 09:06:06 crc kubenswrapper[4873]: I1201 09:06:06.416693 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-46xcn" Dec 01 09:06:07 crc kubenswrapper[4873]: I1201 09:06:07.389175 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-46xcn" Dec 01 09:06:07 crc kubenswrapper[4873]: I1201 09:06:07.449235 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-46xcn"] Dec 01 09:06:09 crc kubenswrapper[4873]: I1201 09:06:09.348566 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-46xcn" podUID="f9aa84f5-b53c-4a95-8204-03128df83d58" containerName="registry-server" containerID="cri-o://b3723557d214bc04f9dc734d9fb0626d6b67632eb283c7c3b65da4bb8f045eb9" gracePeriod=2 Dec 01 09:06:09 crc kubenswrapper[4873]: I1201 09:06:09.854934 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-46xcn" Dec 01 09:06:09 crc kubenswrapper[4873]: I1201 09:06:09.967146 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w5xjn\" (UniqueName: \"kubernetes.io/projected/f9aa84f5-b53c-4a95-8204-03128df83d58-kube-api-access-w5xjn\") pod \"f9aa84f5-b53c-4a95-8204-03128df83d58\" (UID: \"f9aa84f5-b53c-4a95-8204-03128df83d58\") " Dec 01 09:06:09 crc kubenswrapper[4873]: I1201 09:06:09.967385 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9aa84f5-b53c-4a95-8204-03128df83d58-catalog-content\") pod \"f9aa84f5-b53c-4a95-8204-03128df83d58\" (UID: \"f9aa84f5-b53c-4a95-8204-03128df83d58\") " Dec 01 09:06:09 crc kubenswrapper[4873]: I1201 09:06:09.967522 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9aa84f5-b53c-4a95-8204-03128df83d58-utilities\") pod \"f9aa84f5-b53c-4a95-8204-03128df83d58\" (UID: \"f9aa84f5-b53c-4a95-8204-03128df83d58\") " Dec 01 09:06:09 crc kubenswrapper[4873]: I1201 09:06:09.968624 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9aa84f5-b53c-4a95-8204-03128df83d58-utilities" (OuterVolumeSpecName: "utilities") pod "f9aa84f5-b53c-4a95-8204-03128df83d58" (UID: "f9aa84f5-b53c-4a95-8204-03128df83d58"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:06:09 crc kubenswrapper[4873]: I1201 09:06:09.986032 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9aa84f5-b53c-4a95-8204-03128df83d58-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f9aa84f5-b53c-4a95-8204-03128df83d58" (UID: "f9aa84f5-b53c-4a95-8204-03128df83d58"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:06:09 crc kubenswrapper[4873]: I1201 09:06:09.987712 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9aa84f5-b53c-4a95-8204-03128df83d58-kube-api-access-w5xjn" (OuterVolumeSpecName: "kube-api-access-w5xjn") pod "f9aa84f5-b53c-4a95-8204-03128df83d58" (UID: "f9aa84f5-b53c-4a95-8204-03128df83d58"). InnerVolumeSpecName "kube-api-access-w5xjn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:06:10 crc kubenswrapper[4873]: I1201 09:06:10.070061 4873 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9aa84f5-b53c-4a95-8204-03128df83d58-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 09:06:10 crc kubenswrapper[4873]: I1201 09:06:10.070114 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w5xjn\" (UniqueName: \"kubernetes.io/projected/f9aa84f5-b53c-4a95-8204-03128df83d58-kube-api-access-w5xjn\") on node \"crc\" DevicePath \"\"" Dec 01 09:06:10 crc kubenswrapper[4873]: I1201 09:06:10.070131 4873 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9aa84f5-b53c-4a95-8204-03128df83d58-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 09:06:10 crc kubenswrapper[4873]: I1201 09:06:10.364619 4873 generic.go:334] "Generic (PLEG): container finished" podID="f9aa84f5-b53c-4a95-8204-03128df83d58" containerID="b3723557d214bc04f9dc734d9fb0626d6b67632eb283c7c3b65da4bb8f045eb9" exitCode=0 Dec 01 09:06:10 crc kubenswrapper[4873]: I1201 09:06:10.364687 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-46xcn" Dec 01 09:06:10 crc kubenswrapper[4873]: I1201 09:06:10.364748 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-46xcn" event={"ID":"f9aa84f5-b53c-4a95-8204-03128df83d58","Type":"ContainerDied","Data":"b3723557d214bc04f9dc734d9fb0626d6b67632eb283c7c3b65da4bb8f045eb9"} Dec 01 09:06:10 crc kubenswrapper[4873]: I1201 09:06:10.364853 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-46xcn" event={"ID":"f9aa84f5-b53c-4a95-8204-03128df83d58","Type":"ContainerDied","Data":"d390b43835fabc37d992ed00f2d905bfdfd5cf7dc057132605641673305607a5"} Dec 01 09:06:10 crc kubenswrapper[4873]: I1201 09:06:10.364891 4873 scope.go:117] "RemoveContainer" containerID="b3723557d214bc04f9dc734d9fb0626d6b67632eb283c7c3b65da4bb8f045eb9" Dec 01 09:06:10 crc kubenswrapper[4873]: I1201 09:06:10.398872 4873 scope.go:117] "RemoveContainer" containerID="a3d2a9684a9f004b15da114e3d120eaa142ea65d58afc36b7052310b36ebe652" Dec 01 09:06:10 crc kubenswrapper[4873]: I1201 09:06:10.401988 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-46xcn"] Dec 01 09:06:10 crc kubenswrapper[4873]: I1201 09:06:10.411658 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-46xcn"] Dec 01 09:06:10 crc kubenswrapper[4873]: I1201 09:06:10.429091 4873 scope.go:117] "RemoveContainer" containerID="0a10f69ef3b1214f2868cb957b4c6f9b1393e665089901b3ea3120c4b3e3ed74" Dec 01 09:06:10 crc kubenswrapper[4873]: I1201 09:06:10.442857 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9aa84f5-b53c-4a95-8204-03128df83d58" path="/var/lib/kubelet/pods/f9aa84f5-b53c-4a95-8204-03128df83d58/volumes" Dec 01 09:06:10 crc kubenswrapper[4873]: I1201 09:06:10.469638 4873 scope.go:117] "RemoveContainer" containerID="b3723557d214bc04f9dc734d9fb0626d6b67632eb283c7c3b65da4bb8f045eb9" Dec 01 09:06:10 crc kubenswrapper[4873]: E1201 09:06:10.470237 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3723557d214bc04f9dc734d9fb0626d6b67632eb283c7c3b65da4bb8f045eb9\": container with ID starting with b3723557d214bc04f9dc734d9fb0626d6b67632eb283c7c3b65da4bb8f045eb9 not found: ID does not exist" containerID="b3723557d214bc04f9dc734d9fb0626d6b67632eb283c7c3b65da4bb8f045eb9" Dec 01 09:06:10 crc kubenswrapper[4873]: I1201 09:06:10.470285 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3723557d214bc04f9dc734d9fb0626d6b67632eb283c7c3b65da4bb8f045eb9"} err="failed to get container status \"b3723557d214bc04f9dc734d9fb0626d6b67632eb283c7c3b65da4bb8f045eb9\": rpc error: code = NotFound desc = could not find container \"b3723557d214bc04f9dc734d9fb0626d6b67632eb283c7c3b65da4bb8f045eb9\": container with ID starting with b3723557d214bc04f9dc734d9fb0626d6b67632eb283c7c3b65da4bb8f045eb9 not found: ID does not exist" Dec 01 09:06:10 crc kubenswrapper[4873]: I1201 09:06:10.470316 4873 scope.go:117] "RemoveContainer" containerID="a3d2a9684a9f004b15da114e3d120eaa142ea65d58afc36b7052310b36ebe652" Dec 01 09:06:10 crc kubenswrapper[4873]: E1201 09:06:10.471263 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3d2a9684a9f004b15da114e3d120eaa142ea65d58afc36b7052310b36ebe652\": container with ID starting with a3d2a9684a9f004b15da114e3d120eaa142ea65d58afc36b7052310b36ebe652 not found: ID does not exist" containerID="a3d2a9684a9f004b15da114e3d120eaa142ea65d58afc36b7052310b36ebe652" Dec 01 09:06:10 crc kubenswrapper[4873]: I1201 09:06:10.471286 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3d2a9684a9f004b15da114e3d120eaa142ea65d58afc36b7052310b36ebe652"} err="failed to get container status \"a3d2a9684a9f004b15da114e3d120eaa142ea65d58afc36b7052310b36ebe652\": rpc error: code = NotFound desc = could not find container \"a3d2a9684a9f004b15da114e3d120eaa142ea65d58afc36b7052310b36ebe652\": container with ID starting with a3d2a9684a9f004b15da114e3d120eaa142ea65d58afc36b7052310b36ebe652 not found: ID does not exist" Dec 01 09:06:10 crc kubenswrapper[4873]: I1201 09:06:10.471301 4873 scope.go:117] "RemoveContainer" containerID="0a10f69ef3b1214f2868cb957b4c6f9b1393e665089901b3ea3120c4b3e3ed74" Dec 01 09:06:10 crc kubenswrapper[4873]: E1201 09:06:10.472119 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a10f69ef3b1214f2868cb957b4c6f9b1393e665089901b3ea3120c4b3e3ed74\": container with ID starting with 0a10f69ef3b1214f2868cb957b4c6f9b1393e665089901b3ea3120c4b3e3ed74 not found: ID does not exist" containerID="0a10f69ef3b1214f2868cb957b4c6f9b1393e665089901b3ea3120c4b3e3ed74" Dec 01 09:06:10 crc kubenswrapper[4873]: I1201 09:06:10.472298 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a10f69ef3b1214f2868cb957b4c6f9b1393e665089901b3ea3120c4b3e3ed74"} err="failed to get container status \"0a10f69ef3b1214f2868cb957b4c6f9b1393e665089901b3ea3120c4b3e3ed74\": rpc error: code = NotFound desc = could not find container \"0a10f69ef3b1214f2868cb957b4c6f9b1393e665089901b3ea3120c4b3e3ed74\": container with ID starting with 0a10f69ef3b1214f2868cb957b4c6f9b1393e665089901b3ea3120c4b3e3ed74 not found: ID does not exist" Dec 01 09:06:13 crc kubenswrapper[4873]: I1201 09:06:13.430189 4873 scope.go:117] "RemoveContainer" containerID="dd1fd5c773938eaca06d208e8759b6376ff65c7cad3dc995d4b7c31e75f098dc" Dec 01 09:06:13 crc kubenswrapper[4873]: E1201 09:06:13.431030 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:06:28 crc kubenswrapper[4873]: I1201 09:06:28.430919 4873 scope.go:117] "RemoveContainer" containerID="dd1fd5c773938eaca06d208e8759b6376ff65c7cad3dc995d4b7c31e75f098dc" Dec 01 09:06:28 crc kubenswrapper[4873]: E1201 09:06:28.432105 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:06:32 crc kubenswrapper[4873]: I1201 09:06:32.876816 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mm6mk"] Dec 01 09:06:32 crc kubenswrapper[4873]: E1201 09:06:32.878261 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9aa84f5-b53c-4a95-8204-03128df83d58" containerName="registry-server" Dec 01 09:06:32 crc kubenswrapper[4873]: I1201 09:06:32.878286 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9aa84f5-b53c-4a95-8204-03128df83d58" containerName="registry-server" Dec 01 09:06:32 crc kubenswrapper[4873]: E1201 09:06:32.878308 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9aa84f5-b53c-4a95-8204-03128df83d58" containerName="extract-content" Dec 01 09:06:32 crc kubenswrapper[4873]: I1201 09:06:32.878316 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9aa84f5-b53c-4a95-8204-03128df83d58" containerName="extract-content" Dec 01 09:06:32 crc kubenswrapper[4873]: E1201 09:06:32.878338 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9aa84f5-b53c-4a95-8204-03128df83d58" containerName="extract-utilities" Dec 01 09:06:32 crc kubenswrapper[4873]: I1201 09:06:32.878346 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9aa84f5-b53c-4a95-8204-03128df83d58" containerName="extract-utilities" Dec 01 09:06:32 crc kubenswrapper[4873]: I1201 09:06:32.878578 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9aa84f5-b53c-4a95-8204-03128df83d58" containerName="registry-server" Dec 01 09:06:32 crc kubenswrapper[4873]: I1201 09:06:32.880367 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mm6mk" Dec 01 09:06:32 crc kubenswrapper[4873]: I1201 09:06:32.902192 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95af0b7f-934d-4eaf-b63c-5e1a6b2101f3-catalog-content\") pod \"redhat-operators-mm6mk\" (UID: \"95af0b7f-934d-4eaf-b63c-5e1a6b2101f3\") " pod="openshift-marketplace/redhat-operators-mm6mk" Dec 01 09:06:32 crc kubenswrapper[4873]: I1201 09:06:32.902256 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mh7tx\" (UniqueName: \"kubernetes.io/projected/95af0b7f-934d-4eaf-b63c-5e1a6b2101f3-kube-api-access-mh7tx\") pod \"redhat-operators-mm6mk\" (UID: \"95af0b7f-934d-4eaf-b63c-5e1a6b2101f3\") " pod="openshift-marketplace/redhat-operators-mm6mk" Dec 01 09:06:32 crc kubenswrapper[4873]: I1201 09:06:32.902312 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95af0b7f-934d-4eaf-b63c-5e1a6b2101f3-utilities\") pod \"redhat-operators-mm6mk\" (UID: \"95af0b7f-934d-4eaf-b63c-5e1a6b2101f3\") " pod="openshift-marketplace/redhat-operators-mm6mk" Dec 01 09:06:32 crc kubenswrapper[4873]: I1201 09:06:32.915923 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mm6mk"] Dec 01 09:06:33 crc kubenswrapper[4873]: I1201 09:06:33.004892 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95af0b7f-934d-4eaf-b63c-5e1a6b2101f3-utilities\") pod \"redhat-operators-mm6mk\" (UID: \"95af0b7f-934d-4eaf-b63c-5e1a6b2101f3\") " pod="openshift-marketplace/redhat-operators-mm6mk" Dec 01 09:06:33 crc kubenswrapper[4873]: I1201 09:06:33.005113 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95af0b7f-934d-4eaf-b63c-5e1a6b2101f3-catalog-content\") pod \"redhat-operators-mm6mk\" (UID: \"95af0b7f-934d-4eaf-b63c-5e1a6b2101f3\") " pod="openshift-marketplace/redhat-operators-mm6mk" Dec 01 09:06:33 crc kubenswrapper[4873]: I1201 09:06:33.005157 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mh7tx\" (UniqueName: \"kubernetes.io/projected/95af0b7f-934d-4eaf-b63c-5e1a6b2101f3-kube-api-access-mh7tx\") pod \"redhat-operators-mm6mk\" (UID: \"95af0b7f-934d-4eaf-b63c-5e1a6b2101f3\") " pod="openshift-marketplace/redhat-operators-mm6mk" Dec 01 09:06:33 crc kubenswrapper[4873]: I1201 09:06:33.005487 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95af0b7f-934d-4eaf-b63c-5e1a6b2101f3-utilities\") pod \"redhat-operators-mm6mk\" (UID: \"95af0b7f-934d-4eaf-b63c-5e1a6b2101f3\") " pod="openshift-marketplace/redhat-operators-mm6mk" Dec 01 09:06:33 crc kubenswrapper[4873]: I1201 09:06:33.005747 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95af0b7f-934d-4eaf-b63c-5e1a6b2101f3-catalog-content\") pod \"redhat-operators-mm6mk\" (UID: \"95af0b7f-934d-4eaf-b63c-5e1a6b2101f3\") " pod="openshift-marketplace/redhat-operators-mm6mk" Dec 01 09:06:33 crc kubenswrapper[4873]: I1201 09:06:33.031682 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mh7tx\" (UniqueName: \"kubernetes.io/projected/95af0b7f-934d-4eaf-b63c-5e1a6b2101f3-kube-api-access-mh7tx\") pod \"redhat-operators-mm6mk\" (UID: \"95af0b7f-934d-4eaf-b63c-5e1a6b2101f3\") " pod="openshift-marketplace/redhat-operators-mm6mk" Dec 01 09:06:33 crc kubenswrapper[4873]: I1201 09:06:33.211532 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mm6mk" Dec 01 09:06:33 crc kubenswrapper[4873]: I1201 09:06:33.717549 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mm6mk"] Dec 01 09:06:34 crc kubenswrapper[4873]: I1201 09:06:34.636861 4873 generic.go:334] "Generic (PLEG): container finished" podID="95af0b7f-934d-4eaf-b63c-5e1a6b2101f3" containerID="5657f2f0bad208842ecb40eb62da42f41bf72d7628670666729dfa7ab366a642" exitCode=0 Dec 01 09:06:34 crc kubenswrapper[4873]: I1201 09:06:34.636927 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mm6mk" event={"ID":"95af0b7f-934d-4eaf-b63c-5e1a6b2101f3","Type":"ContainerDied","Data":"5657f2f0bad208842ecb40eb62da42f41bf72d7628670666729dfa7ab366a642"} Dec 01 09:06:34 crc kubenswrapper[4873]: I1201 09:06:34.636969 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mm6mk" event={"ID":"95af0b7f-934d-4eaf-b63c-5e1a6b2101f3","Type":"ContainerStarted","Data":"23caa9821dda353aab758fcd6ac88b60226f28d419a293cbb0bb803d80adc0cf"} Dec 01 09:06:35 crc kubenswrapper[4873]: I1201 09:06:35.652395 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mm6mk" event={"ID":"95af0b7f-934d-4eaf-b63c-5e1a6b2101f3","Type":"ContainerStarted","Data":"45dd4a2e9ef3e533528e525385402cc0c43b478549e7199f9e32a5402432ef0c"} Dec 01 09:06:38 crc kubenswrapper[4873]: I1201 09:06:38.251173 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wdpht"] Dec 01 09:06:38 crc kubenswrapper[4873]: I1201 09:06:38.255207 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wdpht" Dec 01 09:06:38 crc kubenswrapper[4873]: I1201 09:06:38.261262 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wdpht"] Dec 01 09:06:38 crc kubenswrapper[4873]: I1201 09:06:38.428983 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2-utilities\") pod \"certified-operators-wdpht\" (UID: \"7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2\") " pod="openshift-marketplace/certified-operators-wdpht" Dec 01 09:06:38 crc kubenswrapper[4873]: I1201 09:06:38.429070 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zz9pk\" (UniqueName: \"kubernetes.io/projected/7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2-kube-api-access-zz9pk\") pod \"certified-operators-wdpht\" (UID: \"7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2\") " pod="openshift-marketplace/certified-operators-wdpht" Dec 01 09:06:38 crc kubenswrapper[4873]: I1201 09:06:38.429107 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2-catalog-content\") pod \"certified-operators-wdpht\" (UID: \"7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2\") " pod="openshift-marketplace/certified-operators-wdpht" Dec 01 09:06:38 crc kubenswrapper[4873]: I1201 09:06:38.532033 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2-utilities\") pod \"certified-operators-wdpht\" (UID: \"7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2\") " pod="openshift-marketplace/certified-operators-wdpht" Dec 01 09:06:38 crc kubenswrapper[4873]: I1201 09:06:38.532101 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zz9pk\" (UniqueName: \"kubernetes.io/projected/7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2-kube-api-access-zz9pk\") pod \"certified-operators-wdpht\" (UID: \"7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2\") " pod="openshift-marketplace/certified-operators-wdpht" Dec 01 09:06:38 crc kubenswrapper[4873]: I1201 09:06:38.532146 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2-catalog-content\") pod \"certified-operators-wdpht\" (UID: \"7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2\") " pod="openshift-marketplace/certified-operators-wdpht" Dec 01 09:06:38 crc kubenswrapper[4873]: I1201 09:06:38.533084 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2-utilities\") pod \"certified-operators-wdpht\" (UID: \"7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2\") " pod="openshift-marketplace/certified-operators-wdpht" Dec 01 09:06:38 crc kubenswrapper[4873]: I1201 09:06:38.533439 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2-catalog-content\") pod \"certified-operators-wdpht\" (UID: \"7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2\") " pod="openshift-marketplace/certified-operators-wdpht" Dec 01 09:06:38 crc kubenswrapper[4873]: I1201 09:06:38.564935 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zz9pk\" (UniqueName: \"kubernetes.io/projected/7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2-kube-api-access-zz9pk\") pod \"certified-operators-wdpht\" (UID: \"7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2\") " pod="openshift-marketplace/certified-operators-wdpht" Dec 01 09:06:38 crc kubenswrapper[4873]: I1201 09:06:38.587091 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wdpht" Dec 01 09:06:38 crc kubenswrapper[4873]: I1201 09:06:38.685938 4873 generic.go:334] "Generic (PLEG): container finished" podID="95af0b7f-934d-4eaf-b63c-5e1a6b2101f3" containerID="45dd4a2e9ef3e533528e525385402cc0c43b478549e7199f9e32a5402432ef0c" exitCode=0 Dec 01 09:06:38 crc kubenswrapper[4873]: I1201 09:06:38.685986 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mm6mk" event={"ID":"95af0b7f-934d-4eaf-b63c-5e1a6b2101f3","Type":"ContainerDied","Data":"45dd4a2e9ef3e533528e525385402cc0c43b478549e7199f9e32a5402432ef0c"} Dec 01 09:06:39 crc kubenswrapper[4873]: I1201 09:06:39.152925 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wdpht"] Dec 01 09:06:39 crc kubenswrapper[4873]: I1201 09:06:39.698230 4873 generic.go:334] "Generic (PLEG): container finished" podID="7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2" containerID="1be5548d97b7d79d7084160d787e015e8c946b9cb8276b15497df3215c4d446a" exitCode=0 Dec 01 09:06:39 crc kubenswrapper[4873]: I1201 09:06:39.698371 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wdpht" event={"ID":"7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2","Type":"ContainerDied","Data":"1be5548d97b7d79d7084160d787e015e8c946b9cb8276b15497df3215c4d446a"} Dec 01 09:06:39 crc kubenswrapper[4873]: I1201 09:06:39.698698 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wdpht" event={"ID":"7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2","Type":"ContainerStarted","Data":"fccd3bc139c33f1b9393f44e321b6cad1f6c1bdab721b2ce1b6ecba87425013b"} Dec 01 09:06:39 crc kubenswrapper[4873]: I1201 09:06:39.702111 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mm6mk" event={"ID":"95af0b7f-934d-4eaf-b63c-5e1a6b2101f3","Type":"ContainerStarted","Data":"ef1a432d2e318e827260ee3123cccc8d2c486baac6fbb6f3aaeef3b139f9150f"} Dec 01 09:06:39 crc kubenswrapper[4873]: I1201 09:06:39.756253 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mm6mk" podStartSLOduration=3.07655757 podStartE2EDuration="7.756225926s" podCreationTimestamp="2025-12-01 09:06:32 +0000 UTC" firstStartedPulling="2025-12-01 09:06:34.639522412 +0000 UTC m=+1570.541630951" lastFinishedPulling="2025-12-01 09:06:39.319190768 +0000 UTC m=+1575.221299307" observedRunningTime="2025-12-01 09:06:39.744253403 +0000 UTC m=+1575.646361942" watchObservedRunningTime="2025-12-01 09:06:39.756225926 +0000 UTC m=+1575.658334465" Dec 01 09:06:41 crc kubenswrapper[4873]: I1201 09:06:41.730068 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wdpht" event={"ID":"7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2","Type":"ContainerStarted","Data":"6031b02ab9ff10e09a006bcbd480a6263760ba560477786e5c5aa11419c24235"} Dec 01 09:06:42 crc kubenswrapper[4873]: I1201 09:06:42.747989 4873 generic.go:334] "Generic (PLEG): container finished" podID="7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2" containerID="6031b02ab9ff10e09a006bcbd480a6263760ba560477786e5c5aa11419c24235" exitCode=0 Dec 01 09:06:42 crc kubenswrapper[4873]: I1201 09:06:42.748209 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wdpht" event={"ID":"7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2","Type":"ContainerDied","Data":"6031b02ab9ff10e09a006bcbd480a6263760ba560477786e5c5aa11419c24235"} Dec 01 09:06:42 crc kubenswrapper[4873]: I1201 09:06:42.751439 4873 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 09:06:43 crc kubenswrapper[4873]: I1201 09:06:43.212261 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mm6mk" Dec 01 09:06:43 crc kubenswrapper[4873]: I1201 09:06:43.212482 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mm6mk" Dec 01 09:06:43 crc kubenswrapper[4873]: I1201 09:06:43.430833 4873 scope.go:117] "RemoveContainer" containerID="dd1fd5c773938eaca06d208e8759b6376ff65c7cad3dc995d4b7c31e75f098dc" Dec 01 09:06:43 crc kubenswrapper[4873]: E1201 09:06:43.431139 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:06:44 crc kubenswrapper[4873]: I1201 09:06:44.270669 4873 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mm6mk" podUID="95af0b7f-934d-4eaf-b63c-5e1a6b2101f3" containerName="registry-server" probeResult="failure" output=< Dec 01 09:06:44 crc kubenswrapper[4873]: timeout: failed to connect service ":50051" within 1s Dec 01 09:06:44 crc kubenswrapper[4873]: > Dec 01 09:06:44 crc kubenswrapper[4873]: I1201 09:06:44.772067 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wdpht" event={"ID":"7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2","Type":"ContainerStarted","Data":"c574e7815cab65128ddb5e4f648bb2dd2b0d0b945788e1d2ca9a9eff7d617abb"} Dec 01 09:06:44 crc kubenswrapper[4873]: I1201 09:06:44.794707 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wdpht" podStartSLOduration=2.740544448 podStartE2EDuration="6.794679813s" podCreationTimestamp="2025-12-01 09:06:38 +0000 UTC" firstStartedPulling="2025-12-01 09:06:39.700592014 +0000 UTC m=+1575.602700553" lastFinishedPulling="2025-12-01 09:06:43.754727369 +0000 UTC m=+1579.656835918" observedRunningTime="2025-12-01 09:06:44.79084596 +0000 UTC m=+1580.692954519" watchObservedRunningTime="2025-12-01 09:06:44.794679813 +0000 UTC m=+1580.696788352" Dec 01 09:06:48 crc kubenswrapper[4873]: I1201 09:06:48.587559 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wdpht" Dec 01 09:06:48 crc kubenswrapper[4873]: I1201 09:06:48.588543 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wdpht" Dec 01 09:06:48 crc kubenswrapper[4873]: I1201 09:06:48.641453 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wdpht" Dec 01 09:06:48 crc kubenswrapper[4873]: I1201 09:06:48.880183 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wdpht" Dec 01 09:06:48 crc kubenswrapper[4873]: I1201 09:06:48.943319 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wdpht"] Dec 01 09:06:50 crc kubenswrapper[4873]: I1201 09:06:50.834097 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wdpht" podUID="7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2" containerName="registry-server" containerID="cri-o://c574e7815cab65128ddb5e4f648bb2dd2b0d0b945788e1d2ca9a9eff7d617abb" gracePeriod=2 Dec 01 09:06:51 crc kubenswrapper[4873]: I1201 09:06:51.292921 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wdpht" Dec 01 09:06:51 crc kubenswrapper[4873]: I1201 09:06:51.355291 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2-utilities\") pod \"7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2\" (UID: \"7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2\") " Dec 01 09:06:51 crc kubenswrapper[4873]: I1201 09:06:51.355455 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zz9pk\" (UniqueName: \"kubernetes.io/projected/7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2-kube-api-access-zz9pk\") pod \"7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2\" (UID: \"7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2\") " Dec 01 09:06:51 crc kubenswrapper[4873]: I1201 09:06:51.355577 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2-catalog-content\") pod \"7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2\" (UID: \"7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2\") " Dec 01 09:06:51 crc kubenswrapper[4873]: I1201 09:06:51.358171 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2-utilities" (OuterVolumeSpecName: "utilities") pod "7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2" (UID: "7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:06:51 crc kubenswrapper[4873]: I1201 09:06:51.379226 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2-kube-api-access-zz9pk" (OuterVolumeSpecName: "kube-api-access-zz9pk") pod "7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2" (UID: "7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2"). InnerVolumeSpecName "kube-api-access-zz9pk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:06:51 crc kubenswrapper[4873]: I1201 09:06:51.406843 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2" (UID: "7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:06:51 crc kubenswrapper[4873]: I1201 09:06:51.471372 4873 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 09:06:51 crc kubenswrapper[4873]: I1201 09:06:51.471414 4873 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 09:06:51 crc kubenswrapper[4873]: I1201 09:06:51.471425 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zz9pk\" (UniqueName: \"kubernetes.io/projected/7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2-kube-api-access-zz9pk\") on node \"crc\" DevicePath \"\"" Dec 01 09:06:51 crc kubenswrapper[4873]: I1201 09:06:51.849348 4873 generic.go:334] "Generic (PLEG): container finished" podID="7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2" containerID="c574e7815cab65128ddb5e4f648bb2dd2b0d0b945788e1d2ca9a9eff7d617abb" exitCode=0 Dec 01 09:06:51 crc kubenswrapper[4873]: I1201 09:06:51.849435 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wdpht" event={"ID":"7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2","Type":"ContainerDied","Data":"c574e7815cab65128ddb5e4f648bb2dd2b0d0b945788e1d2ca9a9eff7d617abb"} Dec 01 09:06:51 crc kubenswrapper[4873]: I1201 09:06:51.849480 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wdpht" event={"ID":"7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2","Type":"ContainerDied","Data":"fccd3bc139c33f1b9393f44e321b6cad1f6c1bdab721b2ce1b6ecba87425013b"} Dec 01 09:06:51 crc kubenswrapper[4873]: I1201 09:06:51.849445 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wdpht" Dec 01 09:06:51 crc kubenswrapper[4873]: I1201 09:06:51.849504 4873 scope.go:117] "RemoveContainer" containerID="c574e7815cab65128ddb5e4f648bb2dd2b0d0b945788e1d2ca9a9eff7d617abb" Dec 01 09:06:51 crc kubenswrapper[4873]: I1201 09:06:51.911281 4873 scope.go:117] "RemoveContainer" containerID="6031b02ab9ff10e09a006bcbd480a6263760ba560477786e5c5aa11419c24235" Dec 01 09:06:51 crc kubenswrapper[4873]: I1201 09:06:51.918119 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wdpht"] Dec 01 09:06:51 crc kubenswrapper[4873]: I1201 09:06:51.942901 4873 scope.go:117] "RemoveContainer" containerID="1be5548d97b7d79d7084160d787e015e8c946b9cb8276b15497df3215c4d446a" Dec 01 09:06:51 crc kubenswrapper[4873]: I1201 09:06:51.975594 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wdpht"] Dec 01 09:06:51 crc kubenswrapper[4873]: I1201 09:06:51.994718 4873 scope.go:117] "RemoveContainer" containerID="c574e7815cab65128ddb5e4f648bb2dd2b0d0b945788e1d2ca9a9eff7d617abb" Dec 01 09:06:51 crc kubenswrapper[4873]: E1201 09:06:51.995507 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c574e7815cab65128ddb5e4f648bb2dd2b0d0b945788e1d2ca9a9eff7d617abb\": container with ID starting with c574e7815cab65128ddb5e4f648bb2dd2b0d0b945788e1d2ca9a9eff7d617abb not found: ID does not exist" containerID="c574e7815cab65128ddb5e4f648bb2dd2b0d0b945788e1d2ca9a9eff7d617abb" Dec 01 09:06:51 crc kubenswrapper[4873]: I1201 09:06:51.995574 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c574e7815cab65128ddb5e4f648bb2dd2b0d0b945788e1d2ca9a9eff7d617abb"} err="failed to get container status \"c574e7815cab65128ddb5e4f648bb2dd2b0d0b945788e1d2ca9a9eff7d617abb\": rpc error: code = NotFound desc = could not find container \"c574e7815cab65128ddb5e4f648bb2dd2b0d0b945788e1d2ca9a9eff7d617abb\": container with ID starting with c574e7815cab65128ddb5e4f648bb2dd2b0d0b945788e1d2ca9a9eff7d617abb not found: ID does not exist" Dec 01 09:06:51 crc kubenswrapper[4873]: I1201 09:06:51.995605 4873 scope.go:117] "RemoveContainer" containerID="6031b02ab9ff10e09a006bcbd480a6263760ba560477786e5c5aa11419c24235" Dec 01 09:06:51 crc kubenswrapper[4873]: E1201 09:06:51.996222 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6031b02ab9ff10e09a006bcbd480a6263760ba560477786e5c5aa11419c24235\": container with ID starting with 6031b02ab9ff10e09a006bcbd480a6263760ba560477786e5c5aa11419c24235 not found: ID does not exist" containerID="6031b02ab9ff10e09a006bcbd480a6263760ba560477786e5c5aa11419c24235" Dec 01 09:06:51 crc kubenswrapper[4873]: I1201 09:06:51.996307 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6031b02ab9ff10e09a006bcbd480a6263760ba560477786e5c5aa11419c24235"} err="failed to get container status \"6031b02ab9ff10e09a006bcbd480a6263760ba560477786e5c5aa11419c24235\": rpc error: code = NotFound desc = could not find container \"6031b02ab9ff10e09a006bcbd480a6263760ba560477786e5c5aa11419c24235\": container with ID starting with 6031b02ab9ff10e09a006bcbd480a6263760ba560477786e5c5aa11419c24235 not found: ID does not exist" Dec 01 09:06:51 crc kubenswrapper[4873]: I1201 09:06:51.996352 4873 scope.go:117] "RemoveContainer" containerID="1be5548d97b7d79d7084160d787e015e8c946b9cb8276b15497df3215c4d446a" Dec 01 09:06:52 crc kubenswrapper[4873]: E1201 09:06:52.000839 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1be5548d97b7d79d7084160d787e015e8c946b9cb8276b15497df3215c4d446a\": container with ID starting with 1be5548d97b7d79d7084160d787e015e8c946b9cb8276b15497df3215c4d446a not found: ID does not exist" containerID="1be5548d97b7d79d7084160d787e015e8c946b9cb8276b15497df3215c4d446a" Dec 01 09:06:52 crc kubenswrapper[4873]: I1201 09:06:52.000918 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1be5548d97b7d79d7084160d787e015e8c946b9cb8276b15497df3215c4d446a"} err="failed to get container status \"1be5548d97b7d79d7084160d787e015e8c946b9cb8276b15497df3215c4d446a\": rpc error: code = NotFound desc = could not find container \"1be5548d97b7d79d7084160d787e015e8c946b9cb8276b15497df3215c4d446a\": container with ID starting with 1be5548d97b7d79d7084160d787e015e8c946b9cb8276b15497df3215c4d446a not found: ID does not exist" Dec 01 09:06:52 crc kubenswrapper[4873]: I1201 09:06:52.444867 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2" path="/var/lib/kubelet/pods/7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2/volumes" Dec 01 09:06:53 crc kubenswrapper[4873]: I1201 09:06:53.283853 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mm6mk" Dec 01 09:06:53 crc kubenswrapper[4873]: I1201 09:06:53.351251 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mm6mk" Dec 01 09:06:54 crc kubenswrapper[4873]: I1201 09:06:54.297265 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mm6mk"] Dec 01 09:06:55 crc kubenswrapper[4873]: I1201 09:06:55.201247 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mm6mk" podUID="95af0b7f-934d-4eaf-b63c-5e1a6b2101f3" containerName="registry-server" containerID="cri-o://ef1a432d2e318e827260ee3123cccc8d2c486baac6fbb6f3aaeef3b139f9150f" gracePeriod=2 Dec 01 09:06:56 crc kubenswrapper[4873]: I1201 09:06:56.226246 4873 generic.go:334] "Generic (PLEG): container finished" podID="95af0b7f-934d-4eaf-b63c-5e1a6b2101f3" containerID="ef1a432d2e318e827260ee3123cccc8d2c486baac6fbb6f3aaeef3b139f9150f" exitCode=0 Dec 01 09:06:56 crc kubenswrapper[4873]: I1201 09:06:56.226318 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mm6mk" event={"ID":"95af0b7f-934d-4eaf-b63c-5e1a6b2101f3","Type":"ContainerDied","Data":"ef1a432d2e318e827260ee3123cccc8d2c486baac6fbb6f3aaeef3b139f9150f"} Dec 01 09:06:56 crc kubenswrapper[4873]: I1201 09:06:56.226786 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mm6mk" event={"ID":"95af0b7f-934d-4eaf-b63c-5e1a6b2101f3","Type":"ContainerDied","Data":"23caa9821dda353aab758fcd6ac88b60226f28d419a293cbb0bb803d80adc0cf"} Dec 01 09:06:56 crc kubenswrapper[4873]: I1201 09:06:56.226838 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="23caa9821dda353aab758fcd6ac88b60226f28d419a293cbb0bb803d80adc0cf" Dec 01 09:06:56 crc kubenswrapper[4873]: I1201 09:06:56.306203 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mm6mk" Dec 01 09:06:56 crc kubenswrapper[4873]: I1201 09:06:56.385190 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mh7tx\" (UniqueName: \"kubernetes.io/projected/95af0b7f-934d-4eaf-b63c-5e1a6b2101f3-kube-api-access-mh7tx\") pod \"95af0b7f-934d-4eaf-b63c-5e1a6b2101f3\" (UID: \"95af0b7f-934d-4eaf-b63c-5e1a6b2101f3\") " Dec 01 09:06:56 crc kubenswrapper[4873]: I1201 09:06:56.385319 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95af0b7f-934d-4eaf-b63c-5e1a6b2101f3-catalog-content\") pod \"95af0b7f-934d-4eaf-b63c-5e1a6b2101f3\" (UID: \"95af0b7f-934d-4eaf-b63c-5e1a6b2101f3\") " Dec 01 09:06:56 crc kubenswrapper[4873]: I1201 09:06:56.385559 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95af0b7f-934d-4eaf-b63c-5e1a6b2101f3-utilities\") pod \"95af0b7f-934d-4eaf-b63c-5e1a6b2101f3\" (UID: \"95af0b7f-934d-4eaf-b63c-5e1a6b2101f3\") " Dec 01 09:06:56 crc kubenswrapper[4873]: I1201 09:06:56.386666 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95af0b7f-934d-4eaf-b63c-5e1a6b2101f3-utilities" (OuterVolumeSpecName: "utilities") pod "95af0b7f-934d-4eaf-b63c-5e1a6b2101f3" (UID: "95af0b7f-934d-4eaf-b63c-5e1a6b2101f3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:06:56 crc kubenswrapper[4873]: I1201 09:06:56.391944 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95af0b7f-934d-4eaf-b63c-5e1a6b2101f3-kube-api-access-mh7tx" (OuterVolumeSpecName: "kube-api-access-mh7tx") pod "95af0b7f-934d-4eaf-b63c-5e1a6b2101f3" (UID: "95af0b7f-934d-4eaf-b63c-5e1a6b2101f3"). InnerVolumeSpecName "kube-api-access-mh7tx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:06:56 crc kubenswrapper[4873]: I1201 09:06:56.430794 4873 scope.go:117] "RemoveContainer" containerID="dd1fd5c773938eaca06d208e8759b6376ff65c7cad3dc995d4b7c31e75f098dc" Dec 01 09:06:56 crc kubenswrapper[4873]: E1201 09:06:56.431190 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:06:56 crc kubenswrapper[4873]: I1201 09:06:56.489445 4873 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95af0b7f-934d-4eaf-b63c-5e1a6b2101f3-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 09:06:56 crc kubenswrapper[4873]: I1201 09:06:56.489715 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mh7tx\" (UniqueName: \"kubernetes.io/projected/95af0b7f-934d-4eaf-b63c-5e1a6b2101f3-kube-api-access-mh7tx\") on node \"crc\" DevicePath \"\"" Dec 01 09:06:56 crc kubenswrapper[4873]: I1201 09:06:56.495702 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95af0b7f-934d-4eaf-b63c-5e1a6b2101f3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "95af0b7f-934d-4eaf-b63c-5e1a6b2101f3" (UID: "95af0b7f-934d-4eaf-b63c-5e1a6b2101f3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:06:56 crc kubenswrapper[4873]: I1201 09:06:56.591075 4873 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95af0b7f-934d-4eaf-b63c-5e1a6b2101f3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 09:06:57 crc kubenswrapper[4873]: I1201 09:06:57.235862 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mm6mk" Dec 01 09:06:57 crc kubenswrapper[4873]: I1201 09:06:57.290192 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mm6mk"] Dec 01 09:06:57 crc kubenswrapper[4873]: I1201 09:06:57.303623 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mm6mk"] Dec 01 09:06:58 crc kubenswrapper[4873]: I1201 09:06:58.443187 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95af0b7f-934d-4eaf-b63c-5e1a6b2101f3" path="/var/lib/kubelet/pods/95af0b7f-934d-4eaf-b63c-5e1a6b2101f3/volumes" Dec 01 09:07:04 crc kubenswrapper[4873]: I1201 09:07:04.642939 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kqqh5"] Dec 01 09:07:04 crc kubenswrapper[4873]: E1201 09:07:04.644167 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2" containerName="extract-content" Dec 01 09:07:04 crc kubenswrapper[4873]: I1201 09:07:04.644189 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2" containerName="extract-content" Dec 01 09:07:04 crc kubenswrapper[4873]: E1201 09:07:04.644209 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95af0b7f-934d-4eaf-b63c-5e1a6b2101f3" containerName="extract-utilities" Dec 01 09:07:04 crc kubenswrapper[4873]: I1201 09:07:04.644216 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="95af0b7f-934d-4eaf-b63c-5e1a6b2101f3" containerName="extract-utilities" Dec 01 09:07:04 crc kubenswrapper[4873]: E1201 09:07:04.644229 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2" containerName="extract-utilities" Dec 01 09:07:04 crc kubenswrapper[4873]: I1201 09:07:04.644235 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2" containerName="extract-utilities" Dec 01 09:07:04 crc kubenswrapper[4873]: E1201 09:07:04.644255 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95af0b7f-934d-4eaf-b63c-5e1a6b2101f3" containerName="registry-server" Dec 01 09:07:04 crc kubenswrapper[4873]: I1201 09:07:04.644261 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="95af0b7f-934d-4eaf-b63c-5e1a6b2101f3" containerName="registry-server" Dec 01 09:07:04 crc kubenswrapper[4873]: E1201 09:07:04.644277 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2" containerName="registry-server" Dec 01 09:07:04 crc kubenswrapper[4873]: I1201 09:07:04.644283 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2" containerName="registry-server" Dec 01 09:07:04 crc kubenswrapper[4873]: E1201 09:07:04.644296 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95af0b7f-934d-4eaf-b63c-5e1a6b2101f3" containerName="extract-content" Dec 01 09:07:04 crc kubenswrapper[4873]: I1201 09:07:04.644302 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="95af0b7f-934d-4eaf-b63c-5e1a6b2101f3" containerName="extract-content" Dec 01 09:07:04 crc kubenswrapper[4873]: I1201 09:07:04.644542 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="7516c203-d8d4-4cf0-bdc6-3596c3f3ebf2" containerName="registry-server" Dec 01 09:07:04 crc kubenswrapper[4873]: I1201 09:07:04.644556 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="95af0b7f-934d-4eaf-b63c-5e1a6b2101f3" containerName="registry-server" Dec 01 09:07:04 crc kubenswrapper[4873]: I1201 09:07:04.646128 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kqqh5" Dec 01 09:07:04 crc kubenswrapper[4873]: I1201 09:07:04.680968 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kqqh5"] Dec 01 09:07:04 crc kubenswrapper[4873]: I1201 09:07:04.791235 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07fcfc4c-2776-4e86-9510-df057e3c088c-utilities\") pod \"community-operators-kqqh5\" (UID: \"07fcfc4c-2776-4e86-9510-df057e3c088c\") " pod="openshift-marketplace/community-operators-kqqh5" Dec 01 09:07:04 crc kubenswrapper[4873]: I1201 09:07:04.791396 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jllwd\" (UniqueName: \"kubernetes.io/projected/07fcfc4c-2776-4e86-9510-df057e3c088c-kube-api-access-jllwd\") pod \"community-operators-kqqh5\" (UID: \"07fcfc4c-2776-4e86-9510-df057e3c088c\") " pod="openshift-marketplace/community-operators-kqqh5" Dec 01 09:07:04 crc kubenswrapper[4873]: I1201 09:07:04.791448 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07fcfc4c-2776-4e86-9510-df057e3c088c-catalog-content\") pod \"community-operators-kqqh5\" (UID: \"07fcfc4c-2776-4e86-9510-df057e3c088c\") " pod="openshift-marketplace/community-operators-kqqh5" Dec 01 09:07:04 crc kubenswrapper[4873]: I1201 09:07:04.893382 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jllwd\" (UniqueName: \"kubernetes.io/projected/07fcfc4c-2776-4e86-9510-df057e3c088c-kube-api-access-jllwd\") pod \"community-operators-kqqh5\" (UID: \"07fcfc4c-2776-4e86-9510-df057e3c088c\") " pod="openshift-marketplace/community-operators-kqqh5" Dec 01 09:07:04 crc kubenswrapper[4873]: I1201 09:07:04.893458 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07fcfc4c-2776-4e86-9510-df057e3c088c-catalog-content\") pod \"community-operators-kqqh5\" (UID: \"07fcfc4c-2776-4e86-9510-df057e3c088c\") " pod="openshift-marketplace/community-operators-kqqh5" Dec 01 09:07:04 crc kubenswrapper[4873]: I1201 09:07:04.893565 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07fcfc4c-2776-4e86-9510-df057e3c088c-utilities\") pod \"community-operators-kqqh5\" (UID: \"07fcfc4c-2776-4e86-9510-df057e3c088c\") " pod="openshift-marketplace/community-operators-kqqh5" Dec 01 09:07:04 crc kubenswrapper[4873]: I1201 09:07:04.894502 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07fcfc4c-2776-4e86-9510-df057e3c088c-utilities\") pod \"community-operators-kqqh5\" (UID: \"07fcfc4c-2776-4e86-9510-df057e3c088c\") " pod="openshift-marketplace/community-operators-kqqh5" Dec 01 09:07:04 crc kubenswrapper[4873]: I1201 09:07:04.894538 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07fcfc4c-2776-4e86-9510-df057e3c088c-catalog-content\") pod \"community-operators-kqqh5\" (UID: \"07fcfc4c-2776-4e86-9510-df057e3c088c\") " pod="openshift-marketplace/community-operators-kqqh5" Dec 01 09:07:04 crc kubenswrapper[4873]: I1201 09:07:04.920280 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jllwd\" (UniqueName: \"kubernetes.io/projected/07fcfc4c-2776-4e86-9510-df057e3c088c-kube-api-access-jllwd\") pod \"community-operators-kqqh5\" (UID: \"07fcfc4c-2776-4e86-9510-df057e3c088c\") " pod="openshift-marketplace/community-operators-kqqh5" Dec 01 09:07:04 crc kubenswrapper[4873]: I1201 09:07:04.980688 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kqqh5" Dec 01 09:07:05 crc kubenswrapper[4873]: I1201 09:07:05.637855 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kqqh5"] Dec 01 09:07:06 crc kubenswrapper[4873]: I1201 09:07:06.339441 4873 generic.go:334] "Generic (PLEG): container finished" podID="07fcfc4c-2776-4e86-9510-df057e3c088c" containerID="aa5fd80afcbfd579739be9f463388b890169f9b3e1129872f55fbff13f866d9d" exitCode=0 Dec 01 09:07:06 crc kubenswrapper[4873]: I1201 09:07:06.339685 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kqqh5" event={"ID":"07fcfc4c-2776-4e86-9510-df057e3c088c","Type":"ContainerDied","Data":"aa5fd80afcbfd579739be9f463388b890169f9b3e1129872f55fbff13f866d9d"} Dec 01 09:07:06 crc kubenswrapper[4873]: I1201 09:07:06.339913 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kqqh5" event={"ID":"07fcfc4c-2776-4e86-9510-df057e3c088c","Type":"ContainerStarted","Data":"acd37a7b0b3dfd364c87b7c713f01e6e451e75f3aadaf75184941ca6a411c925"} Dec 01 09:07:08 crc kubenswrapper[4873]: I1201 09:07:08.367388 4873 generic.go:334] "Generic (PLEG): container finished" podID="07fcfc4c-2776-4e86-9510-df057e3c088c" containerID="33b67b1651a79bae5466e241c7e1cef3e50cc4ce7553f2a2401ba1e29fccc5fc" exitCode=0 Dec 01 09:07:08 crc kubenswrapper[4873]: I1201 09:07:08.367506 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kqqh5" event={"ID":"07fcfc4c-2776-4e86-9510-df057e3c088c","Type":"ContainerDied","Data":"33b67b1651a79bae5466e241c7e1cef3e50cc4ce7553f2a2401ba1e29fccc5fc"} Dec 01 09:07:10 crc kubenswrapper[4873]: I1201 09:07:10.391284 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kqqh5" event={"ID":"07fcfc4c-2776-4e86-9510-df057e3c088c","Type":"ContainerStarted","Data":"dad9355cc565322521d31cc3b6aab4d0f2feeb43eb9917839b97e99520aa93b6"} Dec 01 09:07:10 crc kubenswrapper[4873]: I1201 09:07:10.424142 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kqqh5" podStartSLOduration=3.409768773 podStartE2EDuration="6.424109996s" podCreationTimestamp="2025-12-01 09:07:04 +0000 UTC" firstStartedPulling="2025-12-01 09:07:06.344067857 +0000 UTC m=+1602.246176396" lastFinishedPulling="2025-12-01 09:07:09.35840908 +0000 UTC m=+1605.260517619" observedRunningTime="2025-12-01 09:07:10.409735894 +0000 UTC m=+1606.311844433" watchObservedRunningTime="2025-12-01 09:07:10.424109996 +0000 UTC m=+1606.326218545" Dec 01 09:07:11 crc kubenswrapper[4873]: I1201 09:07:11.430568 4873 scope.go:117] "RemoveContainer" containerID="dd1fd5c773938eaca06d208e8759b6376ff65c7cad3dc995d4b7c31e75f098dc" Dec 01 09:07:11 crc kubenswrapper[4873]: E1201 09:07:11.430881 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:07:14 crc kubenswrapper[4873]: I1201 09:07:14.981420 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kqqh5" Dec 01 09:07:14 crc kubenswrapper[4873]: I1201 09:07:14.981481 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kqqh5" Dec 01 09:07:15 crc kubenswrapper[4873]: I1201 09:07:15.035535 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kqqh5" Dec 01 09:07:15 crc kubenswrapper[4873]: I1201 09:07:15.510963 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kqqh5" Dec 01 09:07:15 crc kubenswrapper[4873]: I1201 09:07:15.582338 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kqqh5"] Dec 01 09:07:17 crc kubenswrapper[4873]: I1201 09:07:17.466195 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-kqqh5" podUID="07fcfc4c-2776-4e86-9510-df057e3c088c" containerName="registry-server" containerID="cri-o://dad9355cc565322521d31cc3b6aab4d0f2feeb43eb9917839b97e99520aa93b6" gracePeriod=2 Dec 01 09:07:19 crc kubenswrapper[4873]: I1201 09:07:19.491293 4873 generic.go:334] "Generic (PLEG): container finished" podID="07fcfc4c-2776-4e86-9510-df057e3c088c" containerID="dad9355cc565322521d31cc3b6aab4d0f2feeb43eb9917839b97e99520aa93b6" exitCode=0 Dec 01 09:07:19 crc kubenswrapper[4873]: I1201 09:07:19.491583 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kqqh5" event={"ID":"07fcfc4c-2776-4e86-9510-df057e3c088c","Type":"ContainerDied","Data":"dad9355cc565322521d31cc3b6aab4d0f2feeb43eb9917839b97e99520aa93b6"} Dec 01 09:07:19 crc kubenswrapper[4873]: I1201 09:07:19.890089 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kqqh5" Dec 01 09:07:19 crc kubenswrapper[4873]: I1201 09:07:19.980740 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07fcfc4c-2776-4e86-9510-df057e3c088c-catalog-content\") pod \"07fcfc4c-2776-4e86-9510-df057e3c088c\" (UID: \"07fcfc4c-2776-4e86-9510-df057e3c088c\") " Dec 01 09:07:19 crc kubenswrapper[4873]: I1201 09:07:19.981163 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07fcfc4c-2776-4e86-9510-df057e3c088c-utilities\") pod \"07fcfc4c-2776-4e86-9510-df057e3c088c\" (UID: \"07fcfc4c-2776-4e86-9510-df057e3c088c\") " Dec 01 09:07:19 crc kubenswrapper[4873]: I1201 09:07:19.981373 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jllwd\" (UniqueName: \"kubernetes.io/projected/07fcfc4c-2776-4e86-9510-df057e3c088c-kube-api-access-jllwd\") pod \"07fcfc4c-2776-4e86-9510-df057e3c088c\" (UID: \"07fcfc4c-2776-4e86-9510-df057e3c088c\") " Dec 01 09:07:19 crc kubenswrapper[4873]: I1201 09:07:19.982076 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07fcfc4c-2776-4e86-9510-df057e3c088c-utilities" (OuterVolumeSpecName: "utilities") pod "07fcfc4c-2776-4e86-9510-df057e3c088c" (UID: "07fcfc4c-2776-4e86-9510-df057e3c088c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:07:19 crc kubenswrapper[4873]: I1201 09:07:19.982329 4873 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07fcfc4c-2776-4e86-9510-df057e3c088c-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 09:07:19 crc kubenswrapper[4873]: I1201 09:07:19.987539 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07fcfc4c-2776-4e86-9510-df057e3c088c-kube-api-access-jllwd" (OuterVolumeSpecName: "kube-api-access-jllwd") pod "07fcfc4c-2776-4e86-9510-df057e3c088c" (UID: "07fcfc4c-2776-4e86-9510-df057e3c088c"). InnerVolumeSpecName "kube-api-access-jllwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:07:20 crc kubenswrapper[4873]: I1201 09:07:20.039326 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07fcfc4c-2776-4e86-9510-df057e3c088c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "07fcfc4c-2776-4e86-9510-df057e3c088c" (UID: "07fcfc4c-2776-4e86-9510-df057e3c088c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:07:20 crc kubenswrapper[4873]: I1201 09:07:20.090044 4873 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07fcfc4c-2776-4e86-9510-df057e3c088c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 09:07:20 crc kubenswrapper[4873]: I1201 09:07:20.090111 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jllwd\" (UniqueName: \"kubernetes.io/projected/07fcfc4c-2776-4e86-9510-df057e3c088c-kube-api-access-jllwd\") on node \"crc\" DevicePath \"\"" Dec 01 09:07:20 crc kubenswrapper[4873]: I1201 09:07:20.507128 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kqqh5" event={"ID":"07fcfc4c-2776-4e86-9510-df057e3c088c","Type":"ContainerDied","Data":"acd37a7b0b3dfd364c87b7c713f01e6e451e75f3aadaf75184941ca6a411c925"} Dec 01 09:07:20 crc kubenswrapper[4873]: I1201 09:07:20.507213 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kqqh5" Dec 01 09:07:20 crc kubenswrapper[4873]: I1201 09:07:20.507715 4873 scope.go:117] "RemoveContainer" containerID="dad9355cc565322521d31cc3b6aab4d0f2feeb43eb9917839b97e99520aa93b6" Dec 01 09:07:20 crc kubenswrapper[4873]: I1201 09:07:20.537872 4873 scope.go:117] "RemoveContainer" containerID="33b67b1651a79bae5466e241c7e1cef3e50cc4ce7553f2a2401ba1e29fccc5fc" Dec 01 09:07:20 crc kubenswrapper[4873]: I1201 09:07:20.538774 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kqqh5"] Dec 01 09:07:20 crc kubenswrapper[4873]: I1201 09:07:20.548436 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-kqqh5"] Dec 01 09:07:20 crc kubenswrapper[4873]: I1201 09:07:20.568980 4873 scope.go:117] "RemoveContainer" containerID="aa5fd80afcbfd579739be9f463388b890169f9b3e1129872f55fbff13f866d9d" Dec 01 09:07:22 crc kubenswrapper[4873]: I1201 09:07:22.478336 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07fcfc4c-2776-4e86-9510-df057e3c088c" path="/var/lib/kubelet/pods/07fcfc4c-2776-4e86-9510-df057e3c088c/volumes" Dec 01 09:07:23 crc kubenswrapper[4873]: I1201 09:07:23.430961 4873 scope.go:117] "RemoveContainer" containerID="dd1fd5c773938eaca06d208e8759b6376ff65c7cad3dc995d4b7c31e75f098dc" Dec 01 09:07:23 crc kubenswrapper[4873]: E1201 09:07:23.431446 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:07:35 crc kubenswrapper[4873]: I1201 09:07:35.429713 4873 scope.go:117] "RemoveContainer" containerID="dd1fd5c773938eaca06d208e8759b6376ff65c7cad3dc995d4b7c31e75f098dc" Dec 01 09:07:35 crc kubenswrapper[4873]: E1201 09:07:35.430792 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:07:39 crc kubenswrapper[4873]: I1201 09:07:39.706508 4873 generic.go:334] "Generic (PLEG): container finished" podID="489b7faf-eb94-459f-b279-7b5cedf53db9" containerID="0b17315b3b408f1a71b69372e87fd41c476a1d6148d2474063bce7ce3f666609" exitCode=0 Dec 01 09:07:39 crc kubenswrapper[4873]: I1201 09:07:39.706613 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-f9cnt" event={"ID":"489b7faf-eb94-459f-b279-7b5cedf53db9","Type":"ContainerDied","Data":"0b17315b3b408f1a71b69372e87fd41c476a1d6148d2474063bce7ce3f666609"} Dec 01 09:07:41 crc kubenswrapper[4873]: I1201 09:07:41.216636 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-f9cnt" Dec 01 09:07:41 crc kubenswrapper[4873]: I1201 09:07:41.289206 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/489b7faf-eb94-459f-b279-7b5cedf53db9-ssh-key\") pod \"489b7faf-eb94-459f-b279-7b5cedf53db9\" (UID: \"489b7faf-eb94-459f-b279-7b5cedf53db9\") " Dec 01 09:07:41 crc kubenswrapper[4873]: I1201 09:07:41.289303 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/489b7faf-eb94-459f-b279-7b5cedf53db9-bootstrap-combined-ca-bundle\") pod \"489b7faf-eb94-459f-b279-7b5cedf53db9\" (UID: \"489b7faf-eb94-459f-b279-7b5cedf53db9\") " Dec 01 09:07:41 crc kubenswrapper[4873]: I1201 09:07:41.289433 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hh22d\" (UniqueName: \"kubernetes.io/projected/489b7faf-eb94-459f-b279-7b5cedf53db9-kube-api-access-hh22d\") pod \"489b7faf-eb94-459f-b279-7b5cedf53db9\" (UID: \"489b7faf-eb94-459f-b279-7b5cedf53db9\") " Dec 01 09:07:41 crc kubenswrapper[4873]: I1201 09:07:41.289485 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/489b7faf-eb94-459f-b279-7b5cedf53db9-inventory\") pod \"489b7faf-eb94-459f-b279-7b5cedf53db9\" (UID: \"489b7faf-eb94-459f-b279-7b5cedf53db9\") " Dec 01 09:07:41 crc kubenswrapper[4873]: I1201 09:07:41.509399 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/489b7faf-eb94-459f-b279-7b5cedf53db9-kube-api-access-hh22d" (OuterVolumeSpecName: "kube-api-access-hh22d") pod "489b7faf-eb94-459f-b279-7b5cedf53db9" (UID: "489b7faf-eb94-459f-b279-7b5cedf53db9"). InnerVolumeSpecName "kube-api-access-hh22d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:07:41 crc kubenswrapper[4873]: I1201 09:07:41.509808 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/489b7faf-eb94-459f-b279-7b5cedf53db9-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "489b7faf-eb94-459f-b279-7b5cedf53db9" (UID: "489b7faf-eb94-459f-b279-7b5cedf53db9"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:07:41 crc kubenswrapper[4873]: I1201 09:07:41.542331 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/489b7faf-eb94-459f-b279-7b5cedf53db9-inventory" (OuterVolumeSpecName: "inventory") pod "489b7faf-eb94-459f-b279-7b5cedf53db9" (UID: "489b7faf-eb94-459f-b279-7b5cedf53db9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:07:41 crc kubenswrapper[4873]: I1201 09:07:41.544291 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/489b7faf-eb94-459f-b279-7b5cedf53db9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "489b7faf-eb94-459f-b279-7b5cedf53db9" (UID: "489b7faf-eb94-459f-b279-7b5cedf53db9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:07:41 crc kubenswrapper[4873]: I1201 09:07:41.581156 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hh22d\" (UniqueName: \"kubernetes.io/projected/489b7faf-eb94-459f-b279-7b5cedf53db9-kube-api-access-hh22d\") on node \"crc\" DevicePath \"\"" Dec 01 09:07:41 crc kubenswrapper[4873]: I1201 09:07:41.581199 4873 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/489b7faf-eb94-459f-b279-7b5cedf53db9-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 09:07:41 crc kubenswrapper[4873]: I1201 09:07:41.581212 4873 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/489b7faf-eb94-459f-b279-7b5cedf53db9-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 09:07:41 crc kubenswrapper[4873]: I1201 09:07:41.581221 4873 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/489b7faf-eb94-459f-b279-7b5cedf53db9-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:07:41 crc kubenswrapper[4873]: I1201 09:07:41.726821 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-f9cnt" event={"ID":"489b7faf-eb94-459f-b279-7b5cedf53db9","Type":"ContainerDied","Data":"adca14a3f7ed588e4e72c1327a3b999ba237a32e2b06efa67686a0d065cecf8e"} Dec 01 09:07:41 crc kubenswrapper[4873]: I1201 09:07:41.726862 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-f9cnt" Dec 01 09:07:41 crc kubenswrapper[4873]: I1201 09:07:41.726887 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="adca14a3f7ed588e4e72c1327a3b999ba237a32e2b06efa67686a0d065cecf8e" Dec 01 09:07:41 crc kubenswrapper[4873]: I1201 09:07:41.850063 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-j2z8p"] Dec 01 09:07:41 crc kubenswrapper[4873]: E1201 09:07:41.851233 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07fcfc4c-2776-4e86-9510-df057e3c088c" containerName="extract-content" Dec 01 09:07:41 crc kubenswrapper[4873]: I1201 09:07:41.851266 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="07fcfc4c-2776-4e86-9510-df057e3c088c" containerName="extract-content" Dec 01 09:07:41 crc kubenswrapper[4873]: E1201 09:07:41.851282 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07fcfc4c-2776-4e86-9510-df057e3c088c" containerName="registry-server" Dec 01 09:07:41 crc kubenswrapper[4873]: I1201 09:07:41.851293 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="07fcfc4c-2776-4e86-9510-df057e3c088c" containerName="registry-server" Dec 01 09:07:41 crc kubenswrapper[4873]: E1201 09:07:41.851346 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="489b7faf-eb94-459f-b279-7b5cedf53db9" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 01 09:07:41 crc kubenswrapper[4873]: I1201 09:07:41.851357 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="489b7faf-eb94-459f-b279-7b5cedf53db9" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 01 09:07:41 crc kubenswrapper[4873]: E1201 09:07:41.851370 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07fcfc4c-2776-4e86-9510-df057e3c088c" containerName="extract-utilities" Dec 01 09:07:41 crc kubenswrapper[4873]: I1201 09:07:41.851379 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="07fcfc4c-2776-4e86-9510-df057e3c088c" containerName="extract-utilities" Dec 01 09:07:41 crc kubenswrapper[4873]: I1201 09:07:41.851642 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="489b7faf-eb94-459f-b279-7b5cedf53db9" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 01 09:07:41 crc kubenswrapper[4873]: I1201 09:07:41.851683 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="07fcfc4c-2776-4e86-9510-df057e3c088c" containerName="registry-server" Dec 01 09:07:41 crc kubenswrapper[4873]: I1201 09:07:41.852737 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-j2z8p" Dec 01 09:07:41 crc kubenswrapper[4873]: I1201 09:07:41.857029 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 09:07:41 crc kubenswrapper[4873]: I1201 09:07:41.857032 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 09:07:41 crc kubenswrapper[4873]: I1201 09:07:41.857527 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9ghqw" Dec 01 09:07:41 crc kubenswrapper[4873]: I1201 09:07:41.861815 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 09:07:41 crc kubenswrapper[4873]: I1201 09:07:41.871551 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-j2z8p"] Dec 01 09:07:41 crc kubenswrapper[4873]: I1201 09:07:41.989166 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1402ba59-201c-4b03-8c0e-478c9d560ef5-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-j2z8p\" (UID: \"1402ba59-201c-4b03-8c0e-478c9d560ef5\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-j2z8p" Dec 01 09:07:41 crc kubenswrapper[4873]: I1201 09:07:41.989241 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7wc8\" (UniqueName: \"kubernetes.io/projected/1402ba59-201c-4b03-8c0e-478c9d560ef5-kube-api-access-k7wc8\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-j2z8p\" (UID: \"1402ba59-201c-4b03-8c0e-478c9d560ef5\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-j2z8p" Dec 01 09:07:41 crc kubenswrapper[4873]: I1201 09:07:41.989648 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1402ba59-201c-4b03-8c0e-478c9d560ef5-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-j2z8p\" (UID: \"1402ba59-201c-4b03-8c0e-478c9d560ef5\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-j2z8p" Dec 01 09:07:42 crc kubenswrapper[4873]: I1201 09:07:42.091597 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1402ba59-201c-4b03-8c0e-478c9d560ef5-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-j2z8p\" (UID: \"1402ba59-201c-4b03-8c0e-478c9d560ef5\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-j2z8p" Dec 01 09:07:42 crc kubenswrapper[4873]: I1201 09:07:42.091786 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1402ba59-201c-4b03-8c0e-478c9d560ef5-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-j2z8p\" (UID: \"1402ba59-201c-4b03-8c0e-478c9d560ef5\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-j2z8p" Dec 01 09:07:42 crc kubenswrapper[4873]: I1201 09:07:42.091827 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7wc8\" (UniqueName: \"kubernetes.io/projected/1402ba59-201c-4b03-8c0e-478c9d560ef5-kube-api-access-k7wc8\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-j2z8p\" (UID: \"1402ba59-201c-4b03-8c0e-478c9d560ef5\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-j2z8p" Dec 01 09:07:42 crc kubenswrapper[4873]: I1201 09:07:42.097664 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1402ba59-201c-4b03-8c0e-478c9d560ef5-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-j2z8p\" (UID: \"1402ba59-201c-4b03-8c0e-478c9d560ef5\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-j2z8p" Dec 01 09:07:42 crc kubenswrapper[4873]: I1201 09:07:42.097663 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1402ba59-201c-4b03-8c0e-478c9d560ef5-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-j2z8p\" (UID: \"1402ba59-201c-4b03-8c0e-478c9d560ef5\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-j2z8p" Dec 01 09:07:42 crc kubenswrapper[4873]: I1201 09:07:42.118668 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7wc8\" (UniqueName: \"kubernetes.io/projected/1402ba59-201c-4b03-8c0e-478c9d560ef5-kube-api-access-k7wc8\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-j2z8p\" (UID: \"1402ba59-201c-4b03-8c0e-478c9d560ef5\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-j2z8p" Dec 01 09:07:42 crc kubenswrapper[4873]: I1201 09:07:42.188040 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-j2z8p" Dec 01 09:07:42 crc kubenswrapper[4873]: I1201 09:07:42.787588 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-j2z8p"] Dec 01 09:07:43 crc kubenswrapper[4873]: I1201 09:07:43.821717 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-j2z8p" event={"ID":"1402ba59-201c-4b03-8c0e-478c9d560ef5","Type":"ContainerStarted","Data":"31d2ccfc633f60cbeb0d8acafbc828bac658b487620bde22884ccb9426f3d64b"} Dec 01 09:07:44 crc kubenswrapper[4873]: I1201 09:07:44.844825 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-j2z8p" event={"ID":"1402ba59-201c-4b03-8c0e-478c9d560ef5","Type":"ContainerStarted","Data":"7a95a64429616277f401a800e029c011e2a05c0a881b88e6a0699f5863f84145"} Dec 01 09:07:44 crc kubenswrapper[4873]: I1201 09:07:44.875815 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-j2z8p" podStartSLOduration=3.230447889 podStartE2EDuration="3.875712913s" podCreationTimestamp="2025-12-01 09:07:41 +0000 UTC" firstStartedPulling="2025-12-01 09:07:42.795550597 +0000 UTC m=+1638.697659136" lastFinishedPulling="2025-12-01 09:07:43.440815631 +0000 UTC m=+1639.342924160" observedRunningTime="2025-12-01 09:07:44.867577914 +0000 UTC m=+1640.769686453" watchObservedRunningTime="2025-12-01 09:07:44.875712913 +0000 UTC m=+1640.777821452" Dec 01 09:07:48 crc kubenswrapper[4873]: I1201 09:07:48.430584 4873 scope.go:117] "RemoveContainer" containerID="dd1fd5c773938eaca06d208e8759b6376ff65c7cad3dc995d4b7c31e75f098dc" Dec 01 09:07:48 crc kubenswrapper[4873]: E1201 09:07:48.431511 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:07:55 crc kubenswrapper[4873]: I1201 09:07:55.548569 4873 scope.go:117] "RemoveContainer" containerID="9d1768aa02375c0c2a6f07207bc68d7fbc7b61220569fa7396770fed5cb31f2c" Dec 01 09:07:55 crc kubenswrapper[4873]: I1201 09:07:55.577297 4873 scope.go:117] "RemoveContainer" containerID="e513eeaccea770ffb3acfe6ab6c5b7450c325a0a3d6dc60e57e7ee633f39ee89" Dec 01 09:08:00 crc kubenswrapper[4873]: I1201 09:08:00.431531 4873 scope.go:117] "RemoveContainer" containerID="dd1fd5c773938eaca06d208e8759b6376ff65c7cad3dc995d4b7c31e75f098dc" Dec 01 09:08:00 crc kubenswrapper[4873]: E1201 09:08:00.434152 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:08:11 crc kubenswrapper[4873]: I1201 09:08:11.430249 4873 scope.go:117] "RemoveContainer" containerID="dd1fd5c773938eaca06d208e8759b6376ff65c7cad3dc995d4b7c31e75f098dc" Dec 01 09:08:11 crc kubenswrapper[4873]: E1201 09:08:11.431316 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:08:22 crc kubenswrapper[4873]: I1201 09:08:22.053506 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-4p8hz"] Dec 01 09:08:22 crc kubenswrapper[4873]: I1201 09:08:22.076560 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-1033-account-create-update-mfx62"] Dec 01 09:08:22 crc kubenswrapper[4873]: I1201 09:08:22.085208 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-23e6-account-create-update-2h4sc"] Dec 01 09:08:22 crc kubenswrapper[4873]: I1201 09:08:22.092339 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-4p8hz"] Dec 01 09:08:22 crc kubenswrapper[4873]: I1201 09:08:22.100630 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-8kwtw"] Dec 01 09:08:22 crc kubenswrapper[4873]: I1201 09:08:22.109568 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-23e6-account-create-update-2h4sc"] Dec 01 09:08:22 crc kubenswrapper[4873]: I1201 09:08:22.118300 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-1033-account-create-update-mfx62"] Dec 01 09:08:22 crc kubenswrapper[4873]: I1201 09:08:22.126568 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-8kwtw"] Dec 01 09:08:22 crc kubenswrapper[4873]: I1201 09:08:22.443334 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00b50e4b-c0ee-4186-95d1-3d793632ceff" path="/var/lib/kubelet/pods/00b50e4b-c0ee-4186-95d1-3d793632ceff/volumes" Dec 01 09:08:22 crc kubenswrapper[4873]: I1201 09:08:22.444831 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ca75802-f944-4266-80b0-3599a0c0311d" path="/var/lib/kubelet/pods/5ca75802-f944-4266-80b0-3599a0c0311d/volumes" Dec 01 09:08:22 crc kubenswrapper[4873]: I1201 09:08:22.445588 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bc6fcca-1230-4aee-aee3-169b34e8c830" path="/var/lib/kubelet/pods/8bc6fcca-1230-4aee-aee3-169b34e8c830/volumes" Dec 01 09:08:22 crc kubenswrapper[4873]: I1201 09:08:22.446409 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d995b8fc-09c8-4ac2-9387-a6ead74e1571" path="/var/lib/kubelet/pods/d995b8fc-09c8-4ac2-9387-a6ead74e1571/volumes" Dec 01 09:08:25 crc kubenswrapper[4873]: I1201 09:08:25.430445 4873 scope.go:117] "RemoveContainer" containerID="dd1fd5c773938eaca06d208e8759b6376ff65c7cad3dc995d4b7c31e75f098dc" Dec 01 09:08:25 crc kubenswrapper[4873]: E1201 09:08:25.431309 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:08:40 crc kubenswrapper[4873]: I1201 09:08:40.430433 4873 scope.go:117] "RemoveContainer" containerID="dd1fd5c773938eaca06d208e8759b6376ff65c7cad3dc995d4b7c31e75f098dc" Dec 01 09:08:40 crc kubenswrapper[4873]: E1201 09:08:40.431998 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:08:43 crc kubenswrapper[4873]: I1201 09:08:43.066572 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-552e-account-create-update-9r7p2"] Dec 01 09:08:43 crc kubenswrapper[4873]: I1201 09:08:43.091814 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-cbzpc"] Dec 01 09:08:43 crc kubenswrapper[4873]: I1201 09:08:43.104416 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-552e-account-create-update-9r7p2"] Dec 01 09:08:43 crc kubenswrapper[4873]: I1201 09:08:43.122375 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-cbzpc"] Dec 01 09:08:44 crc kubenswrapper[4873]: I1201 09:08:44.454397 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98acc7b6-efd7-49b8-bcd5-ae1d67848dd7" path="/var/lib/kubelet/pods/98acc7b6-efd7-49b8-bcd5-ae1d67848dd7/volumes" Dec 01 09:08:44 crc kubenswrapper[4873]: I1201 09:08:44.458541 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d1118a3-3429-457f-bfcf-40e9f173798e" path="/var/lib/kubelet/pods/9d1118a3-3429-457f-bfcf-40e9f173798e/volumes" Dec 01 09:08:51 crc kubenswrapper[4873]: I1201 09:08:51.550601 4873 generic.go:334] "Generic (PLEG): container finished" podID="1402ba59-201c-4b03-8c0e-478c9d560ef5" containerID="7a95a64429616277f401a800e029c011e2a05c0a881b88e6a0699f5863f84145" exitCode=0 Dec 01 09:08:51 crc kubenswrapper[4873]: I1201 09:08:51.550746 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-j2z8p" event={"ID":"1402ba59-201c-4b03-8c0e-478c9d560ef5","Type":"ContainerDied","Data":"7a95a64429616277f401a800e029c011e2a05c0a881b88e6a0699f5863f84145"} Dec 01 09:08:53 crc kubenswrapper[4873]: I1201 09:08:53.005404 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-j2z8p" Dec 01 09:08:53 crc kubenswrapper[4873]: I1201 09:08:53.142591 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1402ba59-201c-4b03-8c0e-478c9d560ef5-inventory\") pod \"1402ba59-201c-4b03-8c0e-478c9d560ef5\" (UID: \"1402ba59-201c-4b03-8c0e-478c9d560ef5\") " Dec 01 09:08:53 crc kubenswrapper[4873]: I1201 09:08:53.142774 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1402ba59-201c-4b03-8c0e-478c9d560ef5-ssh-key\") pod \"1402ba59-201c-4b03-8c0e-478c9d560ef5\" (UID: \"1402ba59-201c-4b03-8c0e-478c9d560ef5\") " Dec 01 09:08:53 crc kubenswrapper[4873]: I1201 09:08:53.142905 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k7wc8\" (UniqueName: \"kubernetes.io/projected/1402ba59-201c-4b03-8c0e-478c9d560ef5-kube-api-access-k7wc8\") pod \"1402ba59-201c-4b03-8c0e-478c9d560ef5\" (UID: \"1402ba59-201c-4b03-8c0e-478c9d560ef5\") " Dec 01 09:08:53 crc kubenswrapper[4873]: I1201 09:08:53.166423 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1402ba59-201c-4b03-8c0e-478c9d560ef5-kube-api-access-k7wc8" (OuterVolumeSpecName: "kube-api-access-k7wc8") pod "1402ba59-201c-4b03-8c0e-478c9d560ef5" (UID: "1402ba59-201c-4b03-8c0e-478c9d560ef5"). InnerVolumeSpecName "kube-api-access-k7wc8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:08:53 crc kubenswrapper[4873]: I1201 09:08:53.217321 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1402ba59-201c-4b03-8c0e-478c9d560ef5-inventory" (OuterVolumeSpecName: "inventory") pod "1402ba59-201c-4b03-8c0e-478c9d560ef5" (UID: "1402ba59-201c-4b03-8c0e-478c9d560ef5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:08:53 crc kubenswrapper[4873]: I1201 09:08:53.241292 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1402ba59-201c-4b03-8c0e-478c9d560ef5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1402ba59-201c-4b03-8c0e-478c9d560ef5" (UID: "1402ba59-201c-4b03-8c0e-478c9d560ef5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:08:53 crc kubenswrapper[4873]: I1201 09:08:53.245938 4873 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1402ba59-201c-4b03-8c0e-478c9d560ef5-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 09:08:53 crc kubenswrapper[4873]: I1201 09:08:53.245987 4873 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1402ba59-201c-4b03-8c0e-478c9d560ef5-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 09:08:53 crc kubenswrapper[4873]: I1201 09:08:53.245997 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k7wc8\" (UniqueName: \"kubernetes.io/projected/1402ba59-201c-4b03-8c0e-478c9d560ef5-kube-api-access-k7wc8\") on node \"crc\" DevicePath \"\"" Dec 01 09:08:53 crc kubenswrapper[4873]: I1201 09:08:53.430480 4873 scope.go:117] "RemoveContainer" containerID="dd1fd5c773938eaca06d208e8759b6376ff65c7cad3dc995d4b7c31e75f098dc" Dec 01 09:08:53 crc kubenswrapper[4873]: E1201 09:08:53.431044 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:08:53 crc kubenswrapper[4873]: I1201 09:08:53.571699 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-j2z8p" event={"ID":"1402ba59-201c-4b03-8c0e-478c9d560ef5","Type":"ContainerDied","Data":"31d2ccfc633f60cbeb0d8acafbc828bac658b487620bde22884ccb9426f3d64b"} Dec 01 09:08:53 crc kubenswrapper[4873]: I1201 09:08:53.572375 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31d2ccfc633f60cbeb0d8acafbc828bac658b487620bde22884ccb9426f3d64b" Dec 01 09:08:53 crc kubenswrapper[4873]: I1201 09:08:53.572461 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-j2z8p" Dec 01 09:08:53 crc kubenswrapper[4873]: I1201 09:08:53.672226 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tsk8x"] Dec 01 09:08:53 crc kubenswrapper[4873]: E1201 09:08:53.672731 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1402ba59-201c-4b03-8c0e-478c9d560ef5" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 01 09:08:53 crc kubenswrapper[4873]: I1201 09:08:53.672754 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="1402ba59-201c-4b03-8c0e-478c9d560ef5" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 01 09:08:53 crc kubenswrapper[4873]: I1201 09:08:53.672969 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="1402ba59-201c-4b03-8c0e-478c9d560ef5" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 01 09:08:53 crc kubenswrapper[4873]: I1201 09:08:53.673854 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tsk8x" Dec 01 09:08:53 crc kubenswrapper[4873]: I1201 09:08:53.676659 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9ghqw" Dec 01 09:08:53 crc kubenswrapper[4873]: I1201 09:08:53.676798 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 09:08:53 crc kubenswrapper[4873]: I1201 09:08:53.677008 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 09:08:53 crc kubenswrapper[4873]: I1201 09:08:53.677184 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 09:08:53 crc kubenswrapper[4873]: I1201 09:08:53.685748 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tsk8x"] Dec 01 09:08:53 crc kubenswrapper[4873]: I1201 09:08:53.755172 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tsk8x\" (UID: \"a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tsk8x" Dec 01 09:08:53 crc kubenswrapper[4873]: I1201 09:08:53.755243 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tsk8x\" (UID: \"a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tsk8x" Dec 01 09:08:53 crc kubenswrapper[4873]: I1201 09:08:53.755382 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lscb9\" (UniqueName: \"kubernetes.io/projected/a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0-kube-api-access-lscb9\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tsk8x\" (UID: \"a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tsk8x" Dec 01 09:08:53 crc kubenswrapper[4873]: I1201 09:08:53.857724 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lscb9\" (UniqueName: \"kubernetes.io/projected/a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0-kube-api-access-lscb9\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tsk8x\" (UID: \"a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tsk8x" Dec 01 09:08:53 crc kubenswrapper[4873]: I1201 09:08:53.857870 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tsk8x\" (UID: \"a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tsk8x" Dec 01 09:08:53 crc kubenswrapper[4873]: I1201 09:08:53.857899 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tsk8x\" (UID: \"a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tsk8x" Dec 01 09:08:53 crc kubenswrapper[4873]: I1201 09:08:53.869192 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tsk8x\" (UID: \"a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tsk8x" Dec 01 09:08:53 crc kubenswrapper[4873]: I1201 09:08:53.872865 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tsk8x\" (UID: \"a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tsk8x" Dec 01 09:08:53 crc kubenswrapper[4873]: I1201 09:08:53.877167 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lscb9\" (UniqueName: \"kubernetes.io/projected/a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0-kube-api-access-lscb9\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tsk8x\" (UID: \"a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tsk8x" Dec 01 09:08:54 crc kubenswrapper[4873]: I1201 09:08:54.004421 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tsk8x" Dec 01 09:08:54 crc kubenswrapper[4873]: I1201 09:08:54.416442 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tsk8x"] Dec 01 09:08:54 crc kubenswrapper[4873]: I1201 09:08:54.586243 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tsk8x" event={"ID":"a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0","Type":"ContainerStarted","Data":"b45007471429a61806118624b3719d418a9eb882c4cd554893339a674eb200df"} Dec 01 09:08:55 crc kubenswrapper[4873]: I1201 09:08:55.652207 4873 scope.go:117] "RemoveContainer" containerID="93054c797b00a9970940d90143c89b4861d57853b5dd9fab3bc751985d960276" Dec 01 09:08:55 crc kubenswrapper[4873]: I1201 09:08:55.744954 4873 scope.go:117] "RemoveContainer" containerID="c71e7e8cd5db62b0e94da0859a60c4145cf609266c32f087c66fe0a385c1874b" Dec 01 09:08:55 crc kubenswrapper[4873]: I1201 09:08:55.775222 4873 scope.go:117] "RemoveContainer" containerID="229af48591ddaf7172ced7c2664e0ea50f54009b4e96c2798791a8ac56514807" Dec 01 09:08:55 crc kubenswrapper[4873]: I1201 09:08:55.826038 4873 scope.go:117] "RemoveContainer" containerID="aba12e9adedc8a0feede329d963df752cb4abb4aac5b21dbfd923e7b47ae44a6" Dec 01 09:08:55 crc kubenswrapper[4873]: I1201 09:08:55.852382 4873 scope.go:117] "RemoveContainer" containerID="9f7e8b7218c9ef0fe039ad1d0baf4dd982f364a3a9363b0368d64452fc74200f" Dec 01 09:08:55 crc kubenswrapper[4873]: I1201 09:08:55.882805 4873 scope.go:117] "RemoveContainer" containerID="924d488dca72675a7c8014797ad15892332b3f100666693db69c08f4fb13fdc0" Dec 01 09:08:56 crc kubenswrapper[4873]: I1201 09:08:56.607426 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tsk8x" event={"ID":"a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0","Type":"ContainerStarted","Data":"9b0a6483fa36f088701b1a9ba0005c29790f19efacef6f133fba1772329d04ca"} Dec 01 09:09:00 crc kubenswrapper[4873]: I1201 09:09:00.650873 4873 generic.go:334] "Generic (PLEG): container finished" podID="a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0" containerID="9b0a6483fa36f088701b1a9ba0005c29790f19efacef6f133fba1772329d04ca" exitCode=0 Dec 01 09:09:00 crc kubenswrapper[4873]: I1201 09:09:00.650960 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tsk8x" event={"ID":"a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0","Type":"ContainerDied","Data":"9b0a6483fa36f088701b1a9ba0005c29790f19efacef6f133fba1772329d04ca"} Dec 01 09:09:02 crc kubenswrapper[4873]: I1201 09:09:02.120964 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tsk8x" Dec 01 09:09:02 crc kubenswrapper[4873]: I1201 09:09:02.267344 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lscb9\" (UniqueName: \"kubernetes.io/projected/a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0-kube-api-access-lscb9\") pod \"a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0\" (UID: \"a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0\") " Dec 01 09:09:02 crc kubenswrapper[4873]: I1201 09:09:02.267547 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0-ssh-key\") pod \"a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0\" (UID: \"a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0\") " Dec 01 09:09:02 crc kubenswrapper[4873]: I1201 09:09:02.267591 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0-inventory\") pod \"a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0\" (UID: \"a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0\") " Dec 01 09:09:02 crc kubenswrapper[4873]: I1201 09:09:02.276067 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0-kube-api-access-lscb9" (OuterVolumeSpecName: "kube-api-access-lscb9") pod "a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0" (UID: "a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0"). InnerVolumeSpecName "kube-api-access-lscb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:09:02 crc kubenswrapper[4873]: I1201 09:09:02.303364 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0" (UID: "a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:09:02 crc kubenswrapper[4873]: I1201 09:09:02.304781 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0-inventory" (OuterVolumeSpecName: "inventory") pod "a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0" (UID: "a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:09:02 crc kubenswrapper[4873]: I1201 09:09:02.370955 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lscb9\" (UniqueName: \"kubernetes.io/projected/a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0-kube-api-access-lscb9\") on node \"crc\" DevicePath \"\"" Dec 01 09:09:02 crc kubenswrapper[4873]: I1201 09:09:02.370992 4873 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 09:09:02 crc kubenswrapper[4873]: I1201 09:09:02.371001 4873 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 09:09:02 crc kubenswrapper[4873]: I1201 09:09:02.677733 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tsk8x" event={"ID":"a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0","Type":"ContainerDied","Data":"b45007471429a61806118624b3719d418a9eb882c4cd554893339a674eb200df"} Dec 01 09:09:02 crc kubenswrapper[4873]: I1201 09:09:02.677787 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b45007471429a61806118624b3719d418a9eb882c4cd554893339a674eb200df" Dec 01 09:09:02 crc kubenswrapper[4873]: I1201 09:09:02.677866 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tsk8x" Dec 01 09:09:02 crc kubenswrapper[4873]: I1201 09:09:02.771718 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-sgwcl"] Dec 01 09:09:02 crc kubenswrapper[4873]: E1201 09:09:02.772480 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 01 09:09:02 crc kubenswrapper[4873]: I1201 09:09:02.772511 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 01 09:09:02 crc kubenswrapper[4873]: I1201 09:09:02.772716 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 01 09:09:02 crc kubenswrapper[4873]: I1201 09:09:02.773731 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sgwcl" Dec 01 09:09:02 crc kubenswrapper[4873]: I1201 09:09:02.777583 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 09:09:02 crc kubenswrapper[4873]: I1201 09:09:02.777842 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 09:09:02 crc kubenswrapper[4873]: I1201 09:09:02.777978 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9ghqw" Dec 01 09:09:02 crc kubenswrapper[4873]: I1201 09:09:02.782533 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 09:09:02 crc kubenswrapper[4873]: I1201 09:09:02.789835 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-sgwcl"] Dec 01 09:09:02 crc kubenswrapper[4873]: I1201 09:09:02.881585 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/829ed711-4da5-47ed-80c2-4af8aff6e84a-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sgwcl\" (UID: \"829ed711-4da5-47ed-80c2-4af8aff6e84a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sgwcl" Dec 01 09:09:02 crc kubenswrapper[4873]: I1201 09:09:02.881691 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mj6dz\" (UniqueName: \"kubernetes.io/projected/829ed711-4da5-47ed-80c2-4af8aff6e84a-kube-api-access-mj6dz\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sgwcl\" (UID: \"829ed711-4da5-47ed-80c2-4af8aff6e84a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sgwcl" Dec 01 09:09:02 crc kubenswrapper[4873]: I1201 09:09:02.881760 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/829ed711-4da5-47ed-80c2-4af8aff6e84a-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sgwcl\" (UID: \"829ed711-4da5-47ed-80c2-4af8aff6e84a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sgwcl" Dec 01 09:09:02 crc kubenswrapper[4873]: I1201 09:09:02.984380 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/829ed711-4da5-47ed-80c2-4af8aff6e84a-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sgwcl\" (UID: \"829ed711-4da5-47ed-80c2-4af8aff6e84a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sgwcl" Dec 01 09:09:02 crc kubenswrapper[4873]: I1201 09:09:02.984578 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/829ed711-4da5-47ed-80c2-4af8aff6e84a-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sgwcl\" (UID: \"829ed711-4da5-47ed-80c2-4af8aff6e84a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sgwcl" Dec 01 09:09:02 crc kubenswrapper[4873]: I1201 09:09:02.984660 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mj6dz\" (UniqueName: \"kubernetes.io/projected/829ed711-4da5-47ed-80c2-4af8aff6e84a-kube-api-access-mj6dz\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sgwcl\" (UID: \"829ed711-4da5-47ed-80c2-4af8aff6e84a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sgwcl" Dec 01 09:09:02 crc kubenswrapper[4873]: I1201 09:09:02.991200 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/829ed711-4da5-47ed-80c2-4af8aff6e84a-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sgwcl\" (UID: \"829ed711-4da5-47ed-80c2-4af8aff6e84a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sgwcl" Dec 01 09:09:02 crc kubenswrapper[4873]: I1201 09:09:02.991633 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/829ed711-4da5-47ed-80c2-4af8aff6e84a-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sgwcl\" (UID: \"829ed711-4da5-47ed-80c2-4af8aff6e84a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sgwcl" Dec 01 09:09:03 crc kubenswrapper[4873]: I1201 09:09:03.006538 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mj6dz\" (UniqueName: \"kubernetes.io/projected/829ed711-4da5-47ed-80c2-4af8aff6e84a-kube-api-access-mj6dz\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sgwcl\" (UID: \"829ed711-4da5-47ed-80c2-4af8aff6e84a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sgwcl" Dec 01 09:09:03 crc kubenswrapper[4873]: I1201 09:09:03.095199 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sgwcl" Dec 01 09:09:03 crc kubenswrapper[4873]: I1201 09:09:03.663923 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-sgwcl"] Dec 01 09:09:03 crc kubenswrapper[4873]: I1201 09:09:03.689416 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sgwcl" event={"ID":"829ed711-4da5-47ed-80c2-4af8aff6e84a","Type":"ContainerStarted","Data":"fc79de969561edff6dd3e1549a96da4163dd19efa99bf3f74c20b3512a8b9ca8"} Dec 01 09:09:04 crc kubenswrapper[4873]: I1201 09:09:04.702468 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sgwcl" event={"ID":"829ed711-4da5-47ed-80c2-4af8aff6e84a","Type":"ContainerStarted","Data":"6a4fdb6abebcf86cf7aaa93b96ee5520f76d39994e852cbb8ec69897ce253fa3"} Dec 01 09:09:04 crc kubenswrapper[4873]: I1201 09:09:04.732411 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sgwcl" podStartSLOduration=2.302598029 podStartE2EDuration="2.732384387s" podCreationTimestamp="2025-12-01 09:09:02 +0000 UTC" firstStartedPulling="2025-12-01 09:09:03.668609748 +0000 UTC m=+1719.570718287" lastFinishedPulling="2025-12-01 09:09:04.098396106 +0000 UTC m=+1720.000504645" observedRunningTime="2025-12-01 09:09:04.718057756 +0000 UTC m=+1720.620166315" watchObservedRunningTime="2025-12-01 09:09:04.732384387 +0000 UTC m=+1720.634492936" Dec 01 09:09:06 crc kubenswrapper[4873]: I1201 09:09:06.067709 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-6k8d4"] Dec 01 09:09:06 crc kubenswrapper[4873]: I1201 09:09:06.077037 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-6k8d4"] Dec 01 09:09:06 crc kubenswrapper[4873]: I1201 09:09:06.444830 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8" path="/var/lib/kubelet/pods/0d17c0cc-fbec-40ae-bd0c-f8e3fc0016b8/volumes" Dec 01 09:09:07 crc kubenswrapper[4873]: I1201 09:09:07.430074 4873 scope.go:117] "RemoveContainer" containerID="dd1fd5c773938eaca06d208e8759b6376ff65c7cad3dc995d4b7c31e75f098dc" Dec 01 09:09:07 crc kubenswrapper[4873]: E1201 09:09:07.430378 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:09:10 crc kubenswrapper[4873]: I1201 09:09:10.053609 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-eb5b-account-create-update-q6hf2"] Dec 01 09:09:10 crc kubenswrapper[4873]: I1201 09:09:10.067594 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-h9jj9"] Dec 01 09:09:10 crc kubenswrapper[4873]: I1201 09:09:10.081983 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-eb5b-account-create-update-q6hf2"] Dec 01 09:09:10 crc kubenswrapper[4873]: I1201 09:09:10.095927 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-h9jj9"] Dec 01 09:09:10 crc kubenswrapper[4873]: I1201 09:09:10.116202 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-m9cfn"] Dec 01 09:09:10 crc kubenswrapper[4873]: I1201 09:09:10.127411 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-m9cfn"] Dec 01 09:09:10 crc kubenswrapper[4873]: I1201 09:09:10.444340 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17c0c1a0-6746-4e83-aa1e-419bc703ac35" path="/var/lib/kubelet/pods/17c0c1a0-6746-4e83-aa1e-419bc703ac35/volumes" Dec 01 09:09:10 crc kubenswrapper[4873]: I1201 09:09:10.445489 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="860ae34b-0993-4a11-ac3a-3774804a0860" path="/var/lib/kubelet/pods/860ae34b-0993-4a11-ac3a-3774804a0860/volumes" Dec 01 09:09:10 crc kubenswrapper[4873]: I1201 09:09:10.446379 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a20240ed-c3f5-444d-854d-718d0f5cbace" path="/var/lib/kubelet/pods/a20240ed-c3f5-444d-854d-718d0f5cbace/volumes" Dec 01 09:09:11 crc kubenswrapper[4873]: I1201 09:09:11.032147 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-3c32-account-create-update-vl8cf"] Dec 01 09:09:11 crc kubenswrapper[4873]: I1201 09:09:11.042263 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-108c-account-create-update-cmrw2"] Dec 01 09:09:11 crc kubenswrapper[4873]: I1201 09:09:11.056543 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-q6bwr"] Dec 01 09:09:11 crc kubenswrapper[4873]: I1201 09:09:11.065211 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-3c32-account-create-update-vl8cf"] Dec 01 09:09:11 crc kubenswrapper[4873]: I1201 09:09:11.072310 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-108c-account-create-update-cmrw2"] Dec 01 09:09:11 crc kubenswrapper[4873]: I1201 09:09:11.079763 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-q6bwr"] Dec 01 09:09:12 crc kubenswrapper[4873]: I1201 09:09:12.443470 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="570a0599-aafa-47fe-8b7a-26056d6a1723" path="/var/lib/kubelet/pods/570a0599-aafa-47fe-8b7a-26056d6a1723/volumes" Dec 01 09:09:12 crc kubenswrapper[4873]: I1201 09:09:12.444665 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69e4256b-7143-4d91-91a0-b4ffdeb36511" path="/var/lib/kubelet/pods/69e4256b-7143-4d91-91a0-b4ffdeb36511/volumes" Dec 01 09:09:12 crc kubenswrapper[4873]: I1201 09:09:12.445583 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5650152-030e-42d2-8db1-b80bfa18bc6f" path="/var/lib/kubelet/pods/b5650152-030e-42d2-8db1-b80bfa18bc6f/volumes" Dec 01 09:09:21 crc kubenswrapper[4873]: I1201 09:09:21.048193 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-wshpl"] Dec 01 09:09:21 crc kubenswrapper[4873]: I1201 09:09:21.071781 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-wshpl"] Dec 01 09:09:22 crc kubenswrapper[4873]: I1201 09:09:22.430023 4873 scope.go:117] "RemoveContainer" containerID="dd1fd5c773938eaca06d208e8759b6376ff65c7cad3dc995d4b7c31e75f098dc" Dec 01 09:09:22 crc kubenswrapper[4873]: E1201 09:09:22.430602 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:09:22 crc kubenswrapper[4873]: I1201 09:09:22.442800 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b86ab79-9562-43e9-a624-f0bc8e1e16c3" path="/var/lib/kubelet/pods/1b86ab79-9562-43e9-a624-f0bc8e1e16c3/volumes" Dec 01 09:09:33 crc kubenswrapper[4873]: I1201 09:09:33.432288 4873 scope.go:117] "RemoveContainer" containerID="dd1fd5c773938eaca06d208e8759b6376ff65c7cad3dc995d4b7c31e75f098dc" Dec 01 09:09:33 crc kubenswrapper[4873]: E1201 09:09:33.433243 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:09:41 crc kubenswrapper[4873]: I1201 09:09:41.134724 4873 generic.go:334] "Generic (PLEG): container finished" podID="829ed711-4da5-47ed-80c2-4af8aff6e84a" containerID="6a4fdb6abebcf86cf7aaa93b96ee5520f76d39994e852cbb8ec69897ce253fa3" exitCode=0 Dec 01 09:09:41 crc kubenswrapper[4873]: I1201 09:09:41.134850 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sgwcl" event={"ID":"829ed711-4da5-47ed-80c2-4af8aff6e84a","Type":"ContainerDied","Data":"6a4fdb6abebcf86cf7aaa93b96ee5520f76d39994e852cbb8ec69897ce253fa3"} Dec 01 09:09:42 crc kubenswrapper[4873]: I1201 09:09:42.673217 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sgwcl" Dec 01 09:09:42 crc kubenswrapper[4873]: I1201 09:09:42.685924 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/829ed711-4da5-47ed-80c2-4af8aff6e84a-inventory\") pod \"829ed711-4da5-47ed-80c2-4af8aff6e84a\" (UID: \"829ed711-4da5-47ed-80c2-4af8aff6e84a\") " Dec 01 09:09:42 crc kubenswrapper[4873]: I1201 09:09:42.686197 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mj6dz\" (UniqueName: \"kubernetes.io/projected/829ed711-4da5-47ed-80c2-4af8aff6e84a-kube-api-access-mj6dz\") pod \"829ed711-4da5-47ed-80c2-4af8aff6e84a\" (UID: \"829ed711-4da5-47ed-80c2-4af8aff6e84a\") " Dec 01 09:09:42 crc kubenswrapper[4873]: I1201 09:09:42.686239 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/829ed711-4da5-47ed-80c2-4af8aff6e84a-ssh-key\") pod \"829ed711-4da5-47ed-80c2-4af8aff6e84a\" (UID: \"829ed711-4da5-47ed-80c2-4af8aff6e84a\") " Dec 01 09:09:42 crc kubenswrapper[4873]: I1201 09:09:42.702098 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/829ed711-4da5-47ed-80c2-4af8aff6e84a-kube-api-access-mj6dz" (OuterVolumeSpecName: "kube-api-access-mj6dz") pod "829ed711-4da5-47ed-80c2-4af8aff6e84a" (UID: "829ed711-4da5-47ed-80c2-4af8aff6e84a"). InnerVolumeSpecName "kube-api-access-mj6dz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:09:42 crc kubenswrapper[4873]: I1201 09:09:42.757876 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/829ed711-4da5-47ed-80c2-4af8aff6e84a-inventory" (OuterVolumeSpecName: "inventory") pod "829ed711-4da5-47ed-80c2-4af8aff6e84a" (UID: "829ed711-4da5-47ed-80c2-4af8aff6e84a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:09:42 crc kubenswrapper[4873]: I1201 09:09:42.775256 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/829ed711-4da5-47ed-80c2-4af8aff6e84a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "829ed711-4da5-47ed-80c2-4af8aff6e84a" (UID: "829ed711-4da5-47ed-80c2-4af8aff6e84a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:09:42 crc kubenswrapper[4873]: I1201 09:09:42.788988 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mj6dz\" (UniqueName: \"kubernetes.io/projected/829ed711-4da5-47ed-80c2-4af8aff6e84a-kube-api-access-mj6dz\") on node \"crc\" DevicePath \"\"" Dec 01 09:09:42 crc kubenswrapper[4873]: I1201 09:09:42.789026 4873 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/829ed711-4da5-47ed-80c2-4af8aff6e84a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 09:09:42 crc kubenswrapper[4873]: I1201 09:09:42.789036 4873 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/829ed711-4da5-47ed-80c2-4af8aff6e84a-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 09:09:43 crc kubenswrapper[4873]: I1201 09:09:43.160058 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sgwcl" event={"ID":"829ed711-4da5-47ed-80c2-4af8aff6e84a","Type":"ContainerDied","Data":"fc79de969561edff6dd3e1549a96da4163dd19efa99bf3f74c20b3512a8b9ca8"} Dec 01 09:09:43 crc kubenswrapper[4873]: I1201 09:09:43.160615 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fc79de969561edff6dd3e1549a96da4163dd19efa99bf3f74c20b3512a8b9ca8" Dec 01 09:09:43 crc kubenswrapper[4873]: I1201 09:09:43.160158 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sgwcl" Dec 01 09:09:43 crc kubenswrapper[4873]: I1201 09:09:43.269403 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ffspk"] Dec 01 09:09:43 crc kubenswrapper[4873]: E1201 09:09:43.270476 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="829ed711-4da5-47ed-80c2-4af8aff6e84a" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 01 09:09:43 crc kubenswrapper[4873]: I1201 09:09:43.270531 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="829ed711-4da5-47ed-80c2-4af8aff6e84a" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 01 09:09:43 crc kubenswrapper[4873]: I1201 09:09:43.271221 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="829ed711-4da5-47ed-80c2-4af8aff6e84a" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 01 09:09:43 crc kubenswrapper[4873]: I1201 09:09:43.272822 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ffspk" Dec 01 09:09:43 crc kubenswrapper[4873]: I1201 09:09:43.277731 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 09:09:43 crc kubenswrapper[4873]: I1201 09:09:43.278007 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 09:09:43 crc kubenswrapper[4873]: I1201 09:09:43.278523 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9ghqw" Dec 01 09:09:43 crc kubenswrapper[4873]: I1201 09:09:43.283324 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ffspk"] Dec 01 09:09:43 crc kubenswrapper[4873]: I1201 09:09:43.293270 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 09:09:43 crc kubenswrapper[4873]: I1201 09:09:43.303545 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ffspk\" (UID: \"2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ffspk" Dec 01 09:09:43 crc kubenswrapper[4873]: I1201 09:09:43.303647 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ffspk\" (UID: \"2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ffspk" Dec 01 09:09:43 crc kubenswrapper[4873]: I1201 09:09:43.303839 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkrjq\" (UniqueName: \"kubernetes.io/projected/2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1-kube-api-access-bkrjq\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ffspk\" (UID: \"2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ffspk" Dec 01 09:09:43 crc kubenswrapper[4873]: I1201 09:09:43.405944 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkrjq\" (UniqueName: \"kubernetes.io/projected/2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1-kube-api-access-bkrjq\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ffspk\" (UID: \"2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ffspk" Dec 01 09:09:43 crc kubenswrapper[4873]: I1201 09:09:43.406189 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ffspk\" (UID: \"2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ffspk" Dec 01 09:09:43 crc kubenswrapper[4873]: I1201 09:09:43.406262 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ffspk\" (UID: \"2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ffspk" Dec 01 09:09:43 crc kubenswrapper[4873]: I1201 09:09:43.412271 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ffspk\" (UID: \"2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ffspk" Dec 01 09:09:43 crc kubenswrapper[4873]: I1201 09:09:43.412854 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ffspk\" (UID: \"2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ffspk" Dec 01 09:09:43 crc kubenswrapper[4873]: I1201 09:09:43.425422 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkrjq\" (UniqueName: \"kubernetes.io/projected/2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1-kube-api-access-bkrjq\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ffspk\" (UID: \"2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ffspk" Dec 01 09:09:43 crc kubenswrapper[4873]: I1201 09:09:43.622992 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ffspk" Dec 01 09:09:44 crc kubenswrapper[4873]: I1201 09:09:44.028103 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ffspk"] Dec 01 09:09:44 crc kubenswrapper[4873]: I1201 09:09:44.170743 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ffspk" event={"ID":"2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1","Type":"ContainerStarted","Data":"668e59fe27508c95976edb0e24e776a7a1771930ebd05b9f14bcda64b01c5579"} Dec 01 09:09:44 crc kubenswrapper[4873]: I1201 09:09:44.430307 4873 scope.go:117] "RemoveContainer" containerID="dd1fd5c773938eaca06d208e8759b6376ff65c7cad3dc995d4b7c31e75f098dc" Dec 01 09:09:44 crc kubenswrapper[4873]: E1201 09:09:44.432562 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:09:45 crc kubenswrapper[4873]: I1201 09:09:45.183634 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ffspk" event={"ID":"2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1","Type":"ContainerStarted","Data":"4f7139adfaa2cc1c1c281536f7a1d5644b92edde4af7aa21aa62b1a7d98fbfc7"} Dec 01 09:09:45 crc kubenswrapper[4873]: I1201 09:09:45.207401 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ffspk" podStartSLOduration=1.536145549 podStartE2EDuration="2.20736823s" podCreationTimestamp="2025-12-01 09:09:43 +0000 UTC" firstStartedPulling="2025-12-01 09:09:44.035698754 +0000 UTC m=+1759.937807293" lastFinishedPulling="2025-12-01 09:09:44.706921435 +0000 UTC m=+1760.609029974" observedRunningTime="2025-12-01 09:09:45.201845545 +0000 UTC m=+1761.103954104" watchObservedRunningTime="2025-12-01 09:09:45.20736823 +0000 UTC m=+1761.109476769" Dec 01 09:09:49 crc kubenswrapper[4873]: I1201 09:09:49.227092 4873 generic.go:334] "Generic (PLEG): container finished" podID="2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1" containerID="4f7139adfaa2cc1c1c281536f7a1d5644b92edde4af7aa21aa62b1a7d98fbfc7" exitCode=0 Dec 01 09:09:49 crc kubenswrapper[4873]: I1201 09:09:49.227199 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ffspk" event={"ID":"2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1","Type":"ContainerDied","Data":"4f7139adfaa2cc1c1c281536f7a1d5644b92edde4af7aa21aa62b1a7d98fbfc7"} Dec 01 09:09:50 crc kubenswrapper[4873]: I1201 09:09:50.042964 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-5vlz5"] Dec 01 09:09:50 crc kubenswrapper[4873]: I1201 09:09:50.055545 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-5vlz5"] Dec 01 09:09:50 crc kubenswrapper[4873]: I1201 09:09:50.467134 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec6a6ffa-2ab6-41c4-b594-17066d11654a" path="/var/lib/kubelet/pods/ec6a6ffa-2ab6-41c4-b594-17066d11654a/volumes" Dec 01 09:09:50 crc kubenswrapper[4873]: I1201 09:09:50.667370 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ffspk" Dec 01 09:09:50 crc kubenswrapper[4873]: I1201 09:09:50.868457 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bkrjq\" (UniqueName: \"kubernetes.io/projected/2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1-kube-api-access-bkrjq\") pod \"2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1\" (UID: \"2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1\") " Dec 01 09:09:50 crc kubenswrapper[4873]: I1201 09:09:50.868690 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1-inventory\") pod \"2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1\" (UID: \"2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1\") " Dec 01 09:09:50 crc kubenswrapper[4873]: I1201 09:09:50.868777 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1-ssh-key\") pod \"2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1\" (UID: \"2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1\") " Dec 01 09:09:50 crc kubenswrapper[4873]: I1201 09:09:50.875309 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1-kube-api-access-bkrjq" (OuterVolumeSpecName: "kube-api-access-bkrjq") pod "2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1" (UID: "2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1"). InnerVolumeSpecName "kube-api-access-bkrjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:09:50 crc kubenswrapper[4873]: I1201 09:09:50.910244 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1" (UID: "2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:09:50 crc kubenswrapper[4873]: I1201 09:09:50.915509 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1-inventory" (OuterVolumeSpecName: "inventory") pod "2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1" (UID: "2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:09:50 crc kubenswrapper[4873]: I1201 09:09:50.971300 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bkrjq\" (UniqueName: \"kubernetes.io/projected/2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1-kube-api-access-bkrjq\") on node \"crc\" DevicePath \"\"" Dec 01 09:09:50 crc kubenswrapper[4873]: I1201 09:09:50.971377 4873 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 09:09:50 crc kubenswrapper[4873]: I1201 09:09:50.971405 4873 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 09:09:51 crc kubenswrapper[4873]: I1201 09:09:51.250884 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ffspk" event={"ID":"2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1","Type":"ContainerDied","Data":"668e59fe27508c95976edb0e24e776a7a1771930ebd05b9f14bcda64b01c5579"} Dec 01 09:09:51 crc kubenswrapper[4873]: I1201 09:09:51.250945 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="668e59fe27508c95976edb0e24e776a7a1771930ebd05b9f14bcda64b01c5579" Dec 01 09:09:51 crc kubenswrapper[4873]: I1201 09:09:51.250998 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ffspk" Dec 01 09:09:51 crc kubenswrapper[4873]: I1201 09:09:51.337473 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dhxtp"] Dec 01 09:09:51 crc kubenswrapper[4873]: E1201 09:09:51.337973 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 01 09:09:51 crc kubenswrapper[4873]: I1201 09:09:51.337997 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 01 09:09:51 crc kubenswrapper[4873]: I1201 09:09:51.338267 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 01 09:09:51 crc kubenswrapper[4873]: I1201 09:09:51.339071 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dhxtp" Dec 01 09:09:51 crc kubenswrapper[4873]: I1201 09:09:51.342427 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9ghqw" Dec 01 09:09:51 crc kubenswrapper[4873]: I1201 09:09:51.343487 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 09:09:51 crc kubenswrapper[4873]: I1201 09:09:51.344284 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 09:09:51 crc kubenswrapper[4873]: I1201 09:09:51.345123 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 09:09:51 crc kubenswrapper[4873]: I1201 09:09:51.350338 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dhxtp"] Dec 01 09:09:51 crc kubenswrapper[4873]: I1201 09:09:51.493854 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c34d2445-0767-413e-a35e-61258e34e344-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dhxtp\" (UID: \"c34d2445-0767-413e-a35e-61258e34e344\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dhxtp" Dec 01 09:09:51 crc kubenswrapper[4873]: I1201 09:09:51.493978 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c34d2445-0767-413e-a35e-61258e34e344-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dhxtp\" (UID: \"c34d2445-0767-413e-a35e-61258e34e344\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dhxtp" Dec 01 09:09:51 crc kubenswrapper[4873]: I1201 09:09:51.494097 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4kzq\" (UniqueName: \"kubernetes.io/projected/c34d2445-0767-413e-a35e-61258e34e344-kube-api-access-v4kzq\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dhxtp\" (UID: \"c34d2445-0767-413e-a35e-61258e34e344\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dhxtp" Dec 01 09:09:51 crc kubenswrapper[4873]: I1201 09:09:51.596045 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c34d2445-0767-413e-a35e-61258e34e344-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dhxtp\" (UID: \"c34d2445-0767-413e-a35e-61258e34e344\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dhxtp" Dec 01 09:09:51 crc kubenswrapper[4873]: I1201 09:09:51.596140 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c34d2445-0767-413e-a35e-61258e34e344-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dhxtp\" (UID: \"c34d2445-0767-413e-a35e-61258e34e344\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dhxtp" Dec 01 09:09:51 crc kubenswrapper[4873]: I1201 09:09:51.596218 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4kzq\" (UniqueName: \"kubernetes.io/projected/c34d2445-0767-413e-a35e-61258e34e344-kube-api-access-v4kzq\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dhxtp\" (UID: \"c34d2445-0767-413e-a35e-61258e34e344\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dhxtp" Dec 01 09:09:51 crc kubenswrapper[4873]: I1201 09:09:51.601702 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c34d2445-0767-413e-a35e-61258e34e344-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dhxtp\" (UID: \"c34d2445-0767-413e-a35e-61258e34e344\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dhxtp" Dec 01 09:09:51 crc kubenswrapper[4873]: I1201 09:09:51.604167 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c34d2445-0767-413e-a35e-61258e34e344-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dhxtp\" (UID: \"c34d2445-0767-413e-a35e-61258e34e344\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dhxtp" Dec 01 09:09:51 crc kubenswrapper[4873]: I1201 09:09:51.615826 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4kzq\" (UniqueName: \"kubernetes.io/projected/c34d2445-0767-413e-a35e-61258e34e344-kube-api-access-v4kzq\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dhxtp\" (UID: \"c34d2445-0767-413e-a35e-61258e34e344\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dhxtp" Dec 01 09:09:51 crc kubenswrapper[4873]: I1201 09:09:51.673334 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dhxtp" Dec 01 09:09:52 crc kubenswrapper[4873]: I1201 09:09:52.283298 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dhxtp"] Dec 01 09:09:53 crc kubenswrapper[4873]: I1201 09:09:53.274160 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dhxtp" event={"ID":"c34d2445-0767-413e-a35e-61258e34e344","Type":"ContainerStarted","Data":"588d5d158144859ea69013dc92fdb06a727a03ad7f05b7ffab29f850c9477eef"} Dec 01 09:09:54 crc kubenswrapper[4873]: I1201 09:09:54.290492 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dhxtp" event={"ID":"c34d2445-0767-413e-a35e-61258e34e344","Type":"ContainerStarted","Data":"754a7b2516f6e400aeaac04d1e0d672b23e9e0fe8b2c778716b9489b5138574e"} Dec 01 09:09:54 crc kubenswrapper[4873]: I1201 09:09:54.312645 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dhxtp" podStartSLOduration=2.653111248 podStartE2EDuration="3.312622882s" podCreationTimestamp="2025-12-01 09:09:51 +0000 UTC" firstStartedPulling="2025-12-01 09:09:52.293650741 +0000 UTC m=+1768.195759280" lastFinishedPulling="2025-12-01 09:09:52.953162375 +0000 UTC m=+1768.855270914" observedRunningTime="2025-12-01 09:09:54.308590173 +0000 UTC m=+1770.210698712" watchObservedRunningTime="2025-12-01 09:09:54.312622882 +0000 UTC m=+1770.214731421" Dec 01 09:09:55 crc kubenswrapper[4873]: I1201 09:09:55.083246 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-nmmcg"] Dec 01 09:09:55 crc kubenswrapper[4873]: I1201 09:09:55.098873 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-kzt5x"] Dec 01 09:09:55 crc kubenswrapper[4873]: I1201 09:09:55.114182 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-nmmcg"] Dec 01 09:09:55 crc kubenswrapper[4873]: I1201 09:09:55.126566 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-kzt5x"] Dec 01 09:09:55 crc kubenswrapper[4873]: I1201 09:09:55.430498 4873 scope.go:117] "RemoveContainer" containerID="dd1fd5c773938eaca06d208e8759b6376ff65c7cad3dc995d4b7c31e75f098dc" Dec 01 09:09:55 crc kubenswrapper[4873]: E1201 09:09:55.430836 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:09:56 crc kubenswrapper[4873]: I1201 09:09:56.049539 4873 scope.go:117] "RemoveContainer" containerID="5a5b3cbb7c950ec13da82b3af05ff052fa137e7cb753076bc7c3b959cf2e22bb" Dec 01 09:09:56 crc kubenswrapper[4873]: I1201 09:09:56.086943 4873 scope.go:117] "RemoveContainer" containerID="df48d3b6b18f16bbf985f991b069677905d37acec7b9825b5ae4eb31fc298878" Dec 01 09:09:56 crc kubenswrapper[4873]: I1201 09:09:56.166454 4873 scope.go:117] "RemoveContainer" containerID="67014367ee4a3300e046b29a2dbfb826636377bb9001a425a2675910de3226fa" Dec 01 09:09:56 crc kubenswrapper[4873]: I1201 09:09:56.198437 4873 scope.go:117] "RemoveContainer" containerID="243d39e2e5e4ccb73cb1c86fc2ff0271389dcbd153fc38dc44f58b96c2a27060" Dec 01 09:09:56 crc kubenswrapper[4873]: I1201 09:09:56.250155 4873 scope.go:117] "RemoveContainer" containerID="08894b35673c62becdf6e89e1d7ceaba3675395a718195b458b0423be85453ac" Dec 01 09:09:56 crc kubenswrapper[4873]: I1201 09:09:56.290326 4873 scope.go:117] "RemoveContainer" containerID="8afd0e65b2766a22e9a567e19eab5c347622e6d24d1fd8a897a5058866896007" Dec 01 09:09:56 crc kubenswrapper[4873]: I1201 09:09:56.352517 4873 scope.go:117] "RemoveContainer" containerID="e81edd13885b27ff4960dd15fb62847773d51dce54133f76e250799efff23ff7" Dec 01 09:09:56 crc kubenswrapper[4873]: I1201 09:09:56.404260 4873 scope.go:117] "RemoveContainer" containerID="d3556f9f6736ecd7a85bdac799730eeae69062e0c3e744ccf474db1e4a374bbd" Dec 01 09:09:56 crc kubenswrapper[4873]: I1201 09:09:56.432805 4873 scope.go:117] "RemoveContainer" containerID="6ac16135d2f590ecdcbb00ff124044df0d52cef64a72cf1df45041ba3f664814" Dec 01 09:09:56 crc kubenswrapper[4873]: I1201 09:09:56.445978 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a34b9de-9fde-4ff2-9881-1dcbea700c28" path="/var/lib/kubelet/pods/0a34b9de-9fde-4ff2-9881-1dcbea700c28/volumes" Dec 01 09:09:56 crc kubenswrapper[4873]: I1201 09:09:56.446753 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0351e0d-24dc-407a-ba64-a38b313cf17c" path="/var/lib/kubelet/pods/b0351e0d-24dc-407a-ba64-a38b313cf17c/volumes" Dec 01 09:09:56 crc kubenswrapper[4873]: I1201 09:09:56.460557 4873 scope.go:117] "RemoveContainer" containerID="14a602dd9e3b3f009e5a5469016a9cbc0c58a301f0f38499ba6035bb682849a6" Dec 01 09:10:07 crc kubenswrapper[4873]: I1201 09:10:07.039078 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-bmhcv"] Dec 01 09:10:07 crc kubenswrapper[4873]: I1201 09:10:07.058802 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-bmhcv"] Dec 01 09:10:07 crc kubenswrapper[4873]: I1201 09:10:07.429990 4873 scope.go:117] "RemoveContainer" containerID="dd1fd5c773938eaca06d208e8759b6376ff65c7cad3dc995d4b7c31e75f098dc" Dec 01 09:10:07 crc kubenswrapper[4873]: E1201 09:10:07.430649 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:10:08 crc kubenswrapper[4873]: I1201 09:10:08.037671 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-9lvtd"] Dec 01 09:10:08 crc kubenswrapper[4873]: I1201 09:10:08.049846 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-9lvtd"] Dec 01 09:10:08 crc kubenswrapper[4873]: I1201 09:10:08.442989 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce" path="/var/lib/kubelet/pods/ba4ef7e6-e389-4e21-af9f-8376c9c4f6ce/volumes" Dec 01 09:10:08 crc kubenswrapper[4873]: I1201 09:10:08.444077 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8c5adca-9738-46b1-9376-01cfb425d839" path="/var/lib/kubelet/pods/c8c5adca-9738-46b1-9376-01cfb425d839/volumes" Dec 01 09:10:20 crc kubenswrapper[4873]: I1201 09:10:20.431279 4873 scope.go:117] "RemoveContainer" containerID="dd1fd5c773938eaca06d208e8759b6376ff65c7cad3dc995d4b7c31e75f098dc" Dec 01 09:10:20 crc kubenswrapper[4873]: E1201 09:10:20.432359 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:10:32 crc kubenswrapper[4873]: I1201 09:10:32.430903 4873 scope.go:117] "RemoveContainer" containerID="dd1fd5c773938eaca06d208e8759b6376ff65c7cad3dc995d4b7c31e75f098dc" Dec 01 09:10:32 crc kubenswrapper[4873]: E1201 09:10:32.431794 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:10:42 crc kubenswrapper[4873]: I1201 09:10:42.804237 4873 generic.go:334] "Generic (PLEG): container finished" podID="c34d2445-0767-413e-a35e-61258e34e344" containerID="754a7b2516f6e400aeaac04d1e0d672b23e9e0fe8b2c778716b9489b5138574e" exitCode=0 Dec 01 09:10:42 crc kubenswrapper[4873]: I1201 09:10:42.804510 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dhxtp" event={"ID":"c34d2445-0767-413e-a35e-61258e34e344","Type":"ContainerDied","Data":"754a7b2516f6e400aeaac04d1e0d672b23e9e0fe8b2c778716b9489b5138574e"} Dec 01 09:10:44 crc kubenswrapper[4873]: I1201 09:10:44.286930 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dhxtp" Dec 01 09:10:44 crc kubenswrapper[4873]: I1201 09:10:44.307006 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c34d2445-0767-413e-a35e-61258e34e344-inventory\") pod \"c34d2445-0767-413e-a35e-61258e34e344\" (UID: \"c34d2445-0767-413e-a35e-61258e34e344\") " Dec 01 09:10:44 crc kubenswrapper[4873]: I1201 09:10:44.311205 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v4kzq\" (UniqueName: \"kubernetes.io/projected/c34d2445-0767-413e-a35e-61258e34e344-kube-api-access-v4kzq\") pod \"c34d2445-0767-413e-a35e-61258e34e344\" (UID: \"c34d2445-0767-413e-a35e-61258e34e344\") " Dec 01 09:10:44 crc kubenswrapper[4873]: I1201 09:10:44.311596 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c34d2445-0767-413e-a35e-61258e34e344-ssh-key\") pod \"c34d2445-0767-413e-a35e-61258e34e344\" (UID: \"c34d2445-0767-413e-a35e-61258e34e344\") " Dec 01 09:10:44 crc kubenswrapper[4873]: I1201 09:10:44.318098 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c34d2445-0767-413e-a35e-61258e34e344-kube-api-access-v4kzq" (OuterVolumeSpecName: "kube-api-access-v4kzq") pod "c34d2445-0767-413e-a35e-61258e34e344" (UID: "c34d2445-0767-413e-a35e-61258e34e344"). InnerVolumeSpecName "kube-api-access-v4kzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:10:44 crc kubenswrapper[4873]: I1201 09:10:44.350147 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c34d2445-0767-413e-a35e-61258e34e344-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c34d2445-0767-413e-a35e-61258e34e344" (UID: "c34d2445-0767-413e-a35e-61258e34e344"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:10:44 crc kubenswrapper[4873]: I1201 09:10:44.353465 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c34d2445-0767-413e-a35e-61258e34e344-inventory" (OuterVolumeSpecName: "inventory") pod "c34d2445-0767-413e-a35e-61258e34e344" (UID: "c34d2445-0767-413e-a35e-61258e34e344"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:10:44 crc kubenswrapper[4873]: I1201 09:10:44.416156 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v4kzq\" (UniqueName: \"kubernetes.io/projected/c34d2445-0767-413e-a35e-61258e34e344-kube-api-access-v4kzq\") on node \"crc\" DevicePath \"\"" Dec 01 09:10:44 crc kubenswrapper[4873]: I1201 09:10:44.416581 4873 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c34d2445-0767-413e-a35e-61258e34e344-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 09:10:44 crc kubenswrapper[4873]: I1201 09:10:44.416593 4873 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c34d2445-0767-413e-a35e-61258e34e344-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 09:10:44 crc kubenswrapper[4873]: I1201 09:10:44.831249 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dhxtp" event={"ID":"c34d2445-0767-413e-a35e-61258e34e344","Type":"ContainerDied","Data":"588d5d158144859ea69013dc92fdb06a727a03ad7f05b7ffab29f850c9477eef"} Dec 01 09:10:44 crc kubenswrapper[4873]: I1201 09:10:44.831307 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="588d5d158144859ea69013dc92fdb06a727a03ad7f05b7ffab29f850c9477eef" Dec 01 09:10:44 crc kubenswrapper[4873]: I1201 09:10:44.831358 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dhxtp" Dec 01 09:10:44 crc kubenswrapper[4873]: I1201 09:10:44.927360 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-5x9zj"] Dec 01 09:10:44 crc kubenswrapper[4873]: E1201 09:10:44.927900 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c34d2445-0767-413e-a35e-61258e34e344" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 01 09:10:44 crc kubenswrapper[4873]: I1201 09:10:44.927927 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="c34d2445-0767-413e-a35e-61258e34e344" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 01 09:10:44 crc kubenswrapper[4873]: I1201 09:10:44.928284 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="c34d2445-0767-413e-a35e-61258e34e344" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 01 09:10:44 crc kubenswrapper[4873]: I1201 09:10:44.929118 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-5x9zj" Dec 01 09:10:44 crc kubenswrapper[4873]: I1201 09:10:44.931738 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9ghqw" Dec 01 09:10:44 crc kubenswrapper[4873]: I1201 09:10:44.931893 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 09:10:44 crc kubenswrapper[4873]: I1201 09:10:44.932344 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 09:10:44 crc kubenswrapper[4873]: I1201 09:10:44.939573 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 09:10:44 crc kubenswrapper[4873]: I1201 09:10:44.950157 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-5x9zj"] Dec 01 09:10:45 crc kubenswrapper[4873]: I1201 09:10:45.028048 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmbs9\" (UniqueName: \"kubernetes.io/projected/8e0f903d-9fb9-4b2b-9194-a9316d8b796b-kube-api-access-xmbs9\") pod \"ssh-known-hosts-edpm-deployment-5x9zj\" (UID: \"8e0f903d-9fb9-4b2b-9194-a9316d8b796b\") " pod="openstack/ssh-known-hosts-edpm-deployment-5x9zj" Dec 01 09:10:45 crc kubenswrapper[4873]: I1201 09:10:45.028174 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8e0f903d-9fb9-4b2b-9194-a9316d8b796b-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-5x9zj\" (UID: \"8e0f903d-9fb9-4b2b-9194-a9316d8b796b\") " pod="openstack/ssh-known-hosts-edpm-deployment-5x9zj" Dec 01 09:10:45 crc kubenswrapper[4873]: I1201 09:10:45.028233 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/8e0f903d-9fb9-4b2b-9194-a9316d8b796b-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-5x9zj\" (UID: \"8e0f903d-9fb9-4b2b-9194-a9316d8b796b\") " pod="openstack/ssh-known-hosts-edpm-deployment-5x9zj" Dec 01 09:10:45 crc kubenswrapper[4873]: I1201 09:10:45.130958 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/8e0f903d-9fb9-4b2b-9194-a9316d8b796b-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-5x9zj\" (UID: \"8e0f903d-9fb9-4b2b-9194-a9316d8b796b\") " pod="openstack/ssh-known-hosts-edpm-deployment-5x9zj" Dec 01 09:10:45 crc kubenswrapper[4873]: I1201 09:10:45.131246 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmbs9\" (UniqueName: \"kubernetes.io/projected/8e0f903d-9fb9-4b2b-9194-a9316d8b796b-kube-api-access-xmbs9\") pod \"ssh-known-hosts-edpm-deployment-5x9zj\" (UID: \"8e0f903d-9fb9-4b2b-9194-a9316d8b796b\") " pod="openstack/ssh-known-hosts-edpm-deployment-5x9zj" Dec 01 09:10:45 crc kubenswrapper[4873]: I1201 09:10:45.131329 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8e0f903d-9fb9-4b2b-9194-a9316d8b796b-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-5x9zj\" (UID: \"8e0f903d-9fb9-4b2b-9194-a9316d8b796b\") " pod="openstack/ssh-known-hosts-edpm-deployment-5x9zj" Dec 01 09:10:45 crc kubenswrapper[4873]: I1201 09:10:45.136496 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/8e0f903d-9fb9-4b2b-9194-a9316d8b796b-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-5x9zj\" (UID: \"8e0f903d-9fb9-4b2b-9194-a9316d8b796b\") " pod="openstack/ssh-known-hosts-edpm-deployment-5x9zj" Dec 01 09:10:45 crc kubenswrapper[4873]: I1201 09:10:45.136814 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8e0f903d-9fb9-4b2b-9194-a9316d8b796b-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-5x9zj\" (UID: \"8e0f903d-9fb9-4b2b-9194-a9316d8b796b\") " pod="openstack/ssh-known-hosts-edpm-deployment-5x9zj" Dec 01 09:10:45 crc kubenswrapper[4873]: I1201 09:10:45.154145 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmbs9\" (UniqueName: \"kubernetes.io/projected/8e0f903d-9fb9-4b2b-9194-a9316d8b796b-kube-api-access-xmbs9\") pod \"ssh-known-hosts-edpm-deployment-5x9zj\" (UID: \"8e0f903d-9fb9-4b2b-9194-a9316d8b796b\") " pod="openstack/ssh-known-hosts-edpm-deployment-5x9zj" Dec 01 09:10:45 crc kubenswrapper[4873]: I1201 09:10:45.254085 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-5x9zj" Dec 01 09:10:46 crc kubenswrapper[4873]: I1201 09:10:45.795973 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-5x9zj"] Dec 01 09:10:46 crc kubenswrapper[4873]: I1201 09:10:45.843070 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-5x9zj" event={"ID":"8e0f903d-9fb9-4b2b-9194-a9316d8b796b","Type":"ContainerStarted","Data":"c0dceaae1829746f428d26ef1d3b02d0e3b9d94dd13f45cc8774c33e45a98c09"} Dec 01 09:10:46 crc kubenswrapper[4873]: I1201 09:10:46.431509 4873 scope.go:117] "RemoveContainer" containerID="dd1fd5c773938eaca06d208e8759b6376ff65c7cad3dc995d4b7c31e75f098dc" Dec 01 09:10:46 crc kubenswrapper[4873]: E1201 09:10:46.431823 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:10:46 crc kubenswrapper[4873]: I1201 09:10:46.856042 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-5x9zj" event={"ID":"8e0f903d-9fb9-4b2b-9194-a9316d8b796b","Type":"ContainerStarted","Data":"0950a0e43d01de59e4c5d634f772f6405eefb78ed37e32db1c3763b4722e6a80"} Dec 01 09:10:46 crc kubenswrapper[4873]: I1201 09:10:46.890153 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-5x9zj" podStartSLOduration=2.226867128 podStartE2EDuration="2.890113713s" podCreationTimestamp="2025-12-01 09:10:44 +0000 UTC" firstStartedPulling="2025-12-01 09:10:45.804112511 +0000 UTC m=+1821.706221050" lastFinishedPulling="2025-12-01 09:10:46.467359096 +0000 UTC m=+1822.369467635" observedRunningTime="2025-12-01 09:10:46.88063639 +0000 UTC m=+1822.782744929" watchObservedRunningTime="2025-12-01 09:10:46.890113713 +0000 UTC m=+1822.792222252" Dec 01 09:10:53 crc kubenswrapper[4873]: I1201 09:10:53.921487 4873 generic.go:334] "Generic (PLEG): container finished" podID="8e0f903d-9fb9-4b2b-9194-a9316d8b796b" containerID="0950a0e43d01de59e4c5d634f772f6405eefb78ed37e32db1c3763b4722e6a80" exitCode=0 Dec 01 09:10:53 crc kubenswrapper[4873]: I1201 09:10:53.921588 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-5x9zj" event={"ID":"8e0f903d-9fb9-4b2b-9194-a9316d8b796b","Type":"ContainerDied","Data":"0950a0e43d01de59e4c5d634f772f6405eefb78ed37e32db1c3763b4722e6a80"} Dec 01 09:10:56 crc kubenswrapper[4873]: I1201 09:10:56.063812 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-5x9zj" Dec 01 09:10:56 crc kubenswrapper[4873]: I1201 09:10:56.252720 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/8e0f903d-9fb9-4b2b-9194-a9316d8b796b-inventory-0\") pod \"8e0f903d-9fb9-4b2b-9194-a9316d8b796b\" (UID: \"8e0f903d-9fb9-4b2b-9194-a9316d8b796b\") " Dec 01 09:10:56 crc kubenswrapper[4873]: I1201 09:10:56.253399 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8e0f903d-9fb9-4b2b-9194-a9316d8b796b-ssh-key-openstack-edpm-ipam\") pod \"8e0f903d-9fb9-4b2b-9194-a9316d8b796b\" (UID: \"8e0f903d-9fb9-4b2b-9194-a9316d8b796b\") " Dec 01 09:10:56 crc kubenswrapper[4873]: I1201 09:10:56.253464 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmbs9\" (UniqueName: \"kubernetes.io/projected/8e0f903d-9fb9-4b2b-9194-a9316d8b796b-kube-api-access-xmbs9\") pod \"8e0f903d-9fb9-4b2b-9194-a9316d8b796b\" (UID: \"8e0f903d-9fb9-4b2b-9194-a9316d8b796b\") " Dec 01 09:10:56 crc kubenswrapper[4873]: I1201 09:10:56.260316 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e0f903d-9fb9-4b2b-9194-a9316d8b796b-kube-api-access-xmbs9" (OuterVolumeSpecName: "kube-api-access-xmbs9") pod "8e0f903d-9fb9-4b2b-9194-a9316d8b796b" (UID: "8e0f903d-9fb9-4b2b-9194-a9316d8b796b"). InnerVolumeSpecName "kube-api-access-xmbs9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:10:56 crc kubenswrapper[4873]: I1201 09:10:56.285479 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e0f903d-9fb9-4b2b-9194-a9316d8b796b-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "8e0f903d-9fb9-4b2b-9194-a9316d8b796b" (UID: "8e0f903d-9fb9-4b2b-9194-a9316d8b796b"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:10:56 crc kubenswrapper[4873]: I1201 09:10:56.288515 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e0f903d-9fb9-4b2b-9194-a9316d8b796b-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "8e0f903d-9fb9-4b2b-9194-a9316d8b796b" (UID: "8e0f903d-9fb9-4b2b-9194-a9316d8b796b"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:10:56 crc kubenswrapper[4873]: I1201 09:10:56.356160 4873 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8e0f903d-9fb9-4b2b-9194-a9316d8b796b-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 01 09:10:56 crc kubenswrapper[4873]: I1201 09:10:56.356206 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmbs9\" (UniqueName: \"kubernetes.io/projected/8e0f903d-9fb9-4b2b-9194-a9316d8b796b-kube-api-access-xmbs9\") on node \"crc\" DevicePath \"\"" Dec 01 09:10:56 crc kubenswrapper[4873]: I1201 09:10:56.356219 4873 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/8e0f903d-9fb9-4b2b-9194-a9316d8b796b-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 01 09:10:56 crc kubenswrapper[4873]: I1201 09:10:56.696279 4873 scope.go:117] "RemoveContainer" containerID="00cb6d0f184dc4c23b88389120c79e311420e96e7e478921739301747dbe7aa5" Dec 01 09:10:56 crc kubenswrapper[4873]: I1201 09:10:56.752784 4873 scope.go:117] "RemoveContainer" containerID="930b7227d8f1722404cdef589bc743c34ec93d72d99ae635ecdc38fc41b85884" Dec 01 09:10:56 crc kubenswrapper[4873]: I1201 09:10:56.788206 4873 scope.go:117] "RemoveContainer" containerID="60ccf9b2d1e68f6660316983c445b3c311c62d81172fe847a0d3d879b5fa6c77" Dec 01 09:10:56 crc kubenswrapper[4873]: I1201 09:10:56.995335 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-5x9zj" event={"ID":"8e0f903d-9fb9-4b2b-9194-a9316d8b796b","Type":"ContainerDied","Data":"c0dceaae1829746f428d26ef1d3b02d0e3b9d94dd13f45cc8774c33e45a98c09"} Dec 01 09:10:56 crc kubenswrapper[4873]: I1201 09:10:56.995377 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c0dceaae1829746f428d26ef1d3b02d0e3b9d94dd13f45cc8774c33e45a98c09" Dec 01 09:10:56 crc kubenswrapper[4873]: I1201 09:10:56.995448 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-5x9zj" Dec 01 09:10:57 crc kubenswrapper[4873]: I1201 09:10:57.155926 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-bbwph"] Dec 01 09:10:57 crc kubenswrapper[4873]: E1201 09:10:57.156457 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e0f903d-9fb9-4b2b-9194-a9316d8b796b" containerName="ssh-known-hosts-edpm-deployment" Dec 01 09:10:57 crc kubenswrapper[4873]: I1201 09:10:57.156473 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e0f903d-9fb9-4b2b-9194-a9316d8b796b" containerName="ssh-known-hosts-edpm-deployment" Dec 01 09:10:57 crc kubenswrapper[4873]: I1201 09:10:57.156668 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e0f903d-9fb9-4b2b-9194-a9316d8b796b" containerName="ssh-known-hosts-edpm-deployment" Dec 01 09:10:57 crc kubenswrapper[4873]: I1201 09:10:57.157377 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-bbwph" Dec 01 09:10:57 crc kubenswrapper[4873]: I1201 09:10:57.160230 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 09:10:57 crc kubenswrapper[4873]: I1201 09:10:57.163440 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 09:10:57 crc kubenswrapper[4873]: I1201 09:10:57.164833 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9ghqw" Dec 01 09:10:57 crc kubenswrapper[4873]: I1201 09:10:57.165061 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 09:10:57 crc kubenswrapper[4873]: I1201 09:10:57.191044 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9f853021-83c9-47c0-bf56-7374ffc662cf-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-bbwph\" (UID: \"9f853021-83c9-47c0-bf56-7374ffc662cf\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-bbwph" Dec 01 09:10:57 crc kubenswrapper[4873]: I1201 09:10:57.191301 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fr9wz\" (UniqueName: \"kubernetes.io/projected/9f853021-83c9-47c0-bf56-7374ffc662cf-kube-api-access-fr9wz\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-bbwph\" (UID: \"9f853021-83c9-47c0-bf56-7374ffc662cf\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-bbwph" Dec 01 09:10:57 crc kubenswrapper[4873]: I1201 09:10:57.191646 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9f853021-83c9-47c0-bf56-7374ffc662cf-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-bbwph\" (UID: \"9f853021-83c9-47c0-bf56-7374ffc662cf\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-bbwph" Dec 01 09:10:57 crc kubenswrapper[4873]: I1201 09:10:57.202086 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-bbwph"] Dec 01 09:10:57 crc kubenswrapper[4873]: I1201 09:10:57.294383 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fr9wz\" (UniqueName: \"kubernetes.io/projected/9f853021-83c9-47c0-bf56-7374ffc662cf-kube-api-access-fr9wz\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-bbwph\" (UID: \"9f853021-83c9-47c0-bf56-7374ffc662cf\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-bbwph" Dec 01 09:10:57 crc kubenswrapper[4873]: I1201 09:10:57.294571 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9f853021-83c9-47c0-bf56-7374ffc662cf-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-bbwph\" (UID: \"9f853021-83c9-47c0-bf56-7374ffc662cf\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-bbwph" Dec 01 09:10:57 crc kubenswrapper[4873]: I1201 09:10:57.294611 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9f853021-83c9-47c0-bf56-7374ffc662cf-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-bbwph\" (UID: \"9f853021-83c9-47c0-bf56-7374ffc662cf\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-bbwph" Dec 01 09:10:57 crc kubenswrapper[4873]: I1201 09:10:57.303003 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9f853021-83c9-47c0-bf56-7374ffc662cf-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-bbwph\" (UID: \"9f853021-83c9-47c0-bf56-7374ffc662cf\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-bbwph" Dec 01 09:10:57 crc kubenswrapper[4873]: I1201 09:10:57.309576 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9f853021-83c9-47c0-bf56-7374ffc662cf-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-bbwph\" (UID: \"9f853021-83c9-47c0-bf56-7374ffc662cf\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-bbwph" Dec 01 09:10:57 crc kubenswrapper[4873]: I1201 09:10:57.313786 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fr9wz\" (UniqueName: \"kubernetes.io/projected/9f853021-83c9-47c0-bf56-7374ffc662cf-kube-api-access-fr9wz\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-bbwph\" (UID: \"9f853021-83c9-47c0-bf56-7374ffc662cf\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-bbwph" Dec 01 09:10:57 crc kubenswrapper[4873]: I1201 09:10:57.495338 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-bbwph" Dec 01 09:10:57 crc kubenswrapper[4873]: I1201 09:10:57.832161 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-bbwph"] Dec 01 09:10:57 crc kubenswrapper[4873]: W1201 09:10:57.842777 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f853021_83c9_47c0_bf56_7374ffc662cf.slice/crio-57c01ee0326725cb84a3eec13fce39dc9b332fea98c14aff741c0b2726b8ae75 WatchSource:0}: Error finding container 57c01ee0326725cb84a3eec13fce39dc9b332fea98c14aff741c0b2726b8ae75: Status 404 returned error can't find the container with id 57c01ee0326725cb84a3eec13fce39dc9b332fea98c14aff741c0b2726b8ae75 Dec 01 09:10:58 crc kubenswrapper[4873]: I1201 09:10:58.010965 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-bbwph" event={"ID":"9f853021-83c9-47c0-bf56-7374ffc662cf","Type":"ContainerStarted","Data":"57c01ee0326725cb84a3eec13fce39dc9b332fea98c14aff741c0b2726b8ae75"} Dec 01 09:10:59 crc kubenswrapper[4873]: I1201 09:10:59.023313 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-bbwph" event={"ID":"9f853021-83c9-47c0-bf56-7374ffc662cf","Type":"ContainerStarted","Data":"e3f8450fc3731b547de2770c41b124219ae667f8edbe702dec2cda9d4e34f94f"} Dec 01 09:10:59 crc kubenswrapper[4873]: I1201 09:10:59.042383 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-bbwph" podStartSLOduration=1.164284034 podStartE2EDuration="2.042352596s" podCreationTimestamp="2025-12-01 09:10:57 +0000 UTC" firstStartedPulling="2025-12-01 09:10:57.845492532 +0000 UTC m=+1833.747601071" lastFinishedPulling="2025-12-01 09:10:58.723561094 +0000 UTC m=+1834.625669633" observedRunningTime="2025-12-01 09:10:59.042306845 +0000 UTC m=+1834.944415404" watchObservedRunningTime="2025-12-01 09:10:59.042352596 +0000 UTC m=+1834.944461145" Dec 01 09:11:00 crc kubenswrapper[4873]: I1201 09:11:00.079088 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-g6w99"] Dec 01 09:11:00 crc kubenswrapper[4873]: I1201 09:11:00.095182 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-w2kdn"] Dec 01 09:11:00 crc kubenswrapper[4873]: I1201 09:11:00.106246 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-g6w99"] Dec 01 09:11:00 crc kubenswrapper[4873]: I1201 09:11:00.112492 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-w2kdn"] Dec 01 09:11:00 crc kubenswrapper[4873]: I1201 09:11:00.121128 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-231a-account-create-update-7vv9t"] Dec 01 09:11:00 crc kubenswrapper[4873]: I1201 09:11:00.130337 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-231a-account-create-update-7vv9t"] Dec 01 09:11:00 crc kubenswrapper[4873]: I1201 09:11:00.443183 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0539778e-b89c-4ae6-8a79-7cb4578a0c1c" path="/var/lib/kubelet/pods/0539778e-b89c-4ae6-8a79-7cb4578a0c1c/volumes" Dec 01 09:11:00 crc kubenswrapper[4873]: I1201 09:11:00.444172 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="951312e8-93e4-42b5-8d28-f5df19c6a2d3" path="/var/lib/kubelet/pods/951312e8-93e4-42b5-8d28-f5df19c6a2d3/volumes" Dec 01 09:11:00 crc kubenswrapper[4873]: I1201 09:11:00.444803 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a21563fd-0cda-4e31-af93-976c1950e56c" path="/var/lib/kubelet/pods/a21563fd-0cda-4e31-af93-976c1950e56c/volumes" Dec 01 09:11:01 crc kubenswrapper[4873]: I1201 09:11:01.041823 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-ld88d"] Dec 01 09:11:01 crc kubenswrapper[4873]: I1201 09:11:01.052500 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-2df0-account-create-update-k6xbx"] Dec 01 09:11:01 crc kubenswrapper[4873]: I1201 09:11:01.061879 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-ld88d"] Dec 01 09:11:01 crc kubenswrapper[4873]: I1201 09:11:01.072347 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-2df0-account-create-update-k6xbx"] Dec 01 09:11:01 crc kubenswrapper[4873]: I1201 09:11:01.081964 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-01d7-account-create-update-rnjjk"] Dec 01 09:11:01 crc kubenswrapper[4873]: I1201 09:11:01.092922 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-01d7-account-create-update-rnjjk"] Dec 01 09:11:01 crc kubenswrapper[4873]: I1201 09:11:01.431469 4873 scope.go:117] "RemoveContainer" containerID="dd1fd5c773938eaca06d208e8759b6376ff65c7cad3dc995d4b7c31e75f098dc" Dec 01 09:11:02 crc kubenswrapper[4873]: I1201 09:11:02.069748 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" event={"ID":"fef7b114-0e07-402d-a37b-315c36011f4b","Type":"ContainerStarted","Data":"281f1071cdd39ef7daaacd7ebb1e90ad778c28e93bebe102e714733c9cf3c602"} Dec 01 09:11:02 crc kubenswrapper[4873]: I1201 09:11:02.446773 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ff3eb73-9732-4b92-a21c-3a195b39f791" path="/var/lib/kubelet/pods/1ff3eb73-9732-4b92-a21c-3a195b39f791/volumes" Dec 01 09:11:02 crc kubenswrapper[4873]: I1201 09:11:02.449094 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76f16e3e-d8e2-4585-aa95-61390b75713b" path="/var/lib/kubelet/pods/76f16e3e-d8e2-4585-aa95-61390b75713b/volumes" Dec 01 09:11:02 crc kubenswrapper[4873]: I1201 09:11:02.449811 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d94ba6ed-6043-486e-bb35-8c14add87f61" path="/var/lib/kubelet/pods/d94ba6ed-6043-486e-bb35-8c14add87f61/volumes" Dec 01 09:11:07 crc kubenswrapper[4873]: I1201 09:11:07.130485 4873 generic.go:334] "Generic (PLEG): container finished" podID="9f853021-83c9-47c0-bf56-7374ffc662cf" containerID="e3f8450fc3731b547de2770c41b124219ae667f8edbe702dec2cda9d4e34f94f" exitCode=0 Dec 01 09:11:07 crc kubenswrapper[4873]: I1201 09:11:07.130613 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-bbwph" event={"ID":"9f853021-83c9-47c0-bf56-7374ffc662cf","Type":"ContainerDied","Data":"e3f8450fc3731b547de2770c41b124219ae667f8edbe702dec2cda9d4e34f94f"} Dec 01 09:11:08 crc kubenswrapper[4873]: I1201 09:11:08.661923 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-bbwph" Dec 01 09:11:08 crc kubenswrapper[4873]: I1201 09:11:08.781272 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9f853021-83c9-47c0-bf56-7374ffc662cf-inventory\") pod \"9f853021-83c9-47c0-bf56-7374ffc662cf\" (UID: \"9f853021-83c9-47c0-bf56-7374ffc662cf\") " Dec 01 09:11:08 crc kubenswrapper[4873]: I1201 09:11:08.781335 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9f853021-83c9-47c0-bf56-7374ffc662cf-ssh-key\") pod \"9f853021-83c9-47c0-bf56-7374ffc662cf\" (UID: \"9f853021-83c9-47c0-bf56-7374ffc662cf\") " Dec 01 09:11:08 crc kubenswrapper[4873]: I1201 09:11:08.781566 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fr9wz\" (UniqueName: \"kubernetes.io/projected/9f853021-83c9-47c0-bf56-7374ffc662cf-kube-api-access-fr9wz\") pod \"9f853021-83c9-47c0-bf56-7374ffc662cf\" (UID: \"9f853021-83c9-47c0-bf56-7374ffc662cf\") " Dec 01 09:11:08 crc kubenswrapper[4873]: I1201 09:11:08.802647 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f853021-83c9-47c0-bf56-7374ffc662cf-kube-api-access-fr9wz" (OuterVolumeSpecName: "kube-api-access-fr9wz") pod "9f853021-83c9-47c0-bf56-7374ffc662cf" (UID: "9f853021-83c9-47c0-bf56-7374ffc662cf"). InnerVolumeSpecName "kube-api-access-fr9wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:11:08 crc kubenswrapper[4873]: I1201 09:11:08.815662 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f853021-83c9-47c0-bf56-7374ffc662cf-inventory" (OuterVolumeSpecName: "inventory") pod "9f853021-83c9-47c0-bf56-7374ffc662cf" (UID: "9f853021-83c9-47c0-bf56-7374ffc662cf"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:11:08 crc kubenswrapper[4873]: I1201 09:11:08.833551 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f853021-83c9-47c0-bf56-7374ffc662cf-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9f853021-83c9-47c0-bf56-7374ffc662cf" (UID: "9f853021-83c9-47c0-bf56-7374ffc662cf"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:11:08 crc kubenswrapper[4873]: I1201 09:11:08.886302 4873 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9f853021-83c9-47c0-bf56-7374ffc662cf-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 09:11:08 crc kubenswrapper[4873]: I1201 09:11:08.886627 4873 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9f853021-83c9-47c0-bf56-7374ffc662cf-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 09:11:08 crc kubenswrapper[4873]: I1201 09:11:08.886732 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fr9wz\" (UniqueName: \"kubernetes.io/projected/9f853021-83c9-47c0-bf56-7374ffc662cf-kube-api-access-fr9wz\") on node \"crc\" DevicePath \"\"" Dec 01 09:11:09 crc kubenswrapper[4873]: I1201 09:11:09.159991 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-bbwph" event={"ID":"9f853021-83c9-47c0-bf56-7374ffc662cf","Type":"ContainerDied","Data":"57c01ee0326725cb84a3eec13fce39dc9b332fea98c14aff741c0b2726b8ae75"} Dec 01 09:11:09 crc kubenswrapper[4873]: I1201 09:11:09.160084 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="57c01ee0326725cb84a3eec13fce39dc9b332fea98c14aff741c0b2726b8ae75" Dec 01 09:11:09 crc kubenswrapper[4873]: I1201 09:11:09.160203 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-bbwph" Dec 01 09:11:09 crc kubenswrapper[4873]: I1201 09:11:09.242924 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2f94z"] Dec 01 09:11:09 crc kubenswrapper[4873]: E1201 09:11:09.243511 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f853021-83c9-47c0-bf56-7374ffc662cf" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 01 09:11:09 crc kubenswrapper[4873]: I1201 09:11:09.243544 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f853021-83c9-47c0-bf56-7374ffc662cf" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 01 09:11:09 crc kubenswrapper[4873]: I1201 09:11:09.243827 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f853021-83c9-47c0-bf56-7374ffc662cf" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 01 09:11:09 crc kubenswrapper[4873]: I1201 09:11:09.244748 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2f94z" Dec 01 09:11:09 crc kubenswrapper[4873]: I1201 09:11:09.249769 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 09:11:09 crc kubenswrapper[4873]: I1201 09:11:09.250158 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 09:11:09 crc kubenswrapper[4873]: I1201 09:11:09.251442 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9ghqw" Dec 01 09:11:09 crc kubenswrapper[4873]: I1201 09:11:09.251816 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 09:11:09 crc kubenswrapper[4873]: I1201 09:11:09.258690 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2f94z"] Dec 01 09:11:09 crc kubenswrapper[4873]: I1201 09:11:09.399736 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxxs9\" (UniqueName: \"kubernetes.io/projected/2aee299e-6010-491d-bff8-81b4ef2a52c5-kube-api-access-mxxs9\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-2f94z\" (UID: \"2aee299e-6010-491d-bff8-81b4ef2a52c5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2f94z" Dec 01 09:11:09 crc kubenswrapper[4873]: I1201 09:11:09.400277 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2aee299e-6010-491d-bff8-81b4ef2a52c5-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-2f94z\" (UID: \"2aee299e-6010-491d-bff8-81b4ef2a52c5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2f94z" Dec 01 09:11:09 crc kubenswrapper[4873]: I1201 09:11:09.400398 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2aee299e-6010-491d-bff8-81b4ef2a52c5-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-2f94z\" (UID: \"2aee299e-6010-491d-bff8-81b4ef2a52c5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2f94z" Dec 01 09:11:09 crc kubenswrapper[4873]: I1201 09:11:09.502676 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2aee299e-6010-491d-bff8-81b4ef2a52c5-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-2f94z\" (UID: \"2aee299e-6010-491d-bff8-81b4ef2a52c5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2f94z" Dec 01 09:11:09 crc kubenswrapper[4873]: I1201 09:11:09.502792 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2aee299e-6010-491d-bff8-81b4ef2a52c5-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-2f94z\" (UID: \"2aee299e-6010-491d-bff8-81b4ef2a52c5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2f94z" Dec 01 09:11:09 crc kubenswrapper[4873]: I1201 09:11:09.502937 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxxs9\" (UniqueName: \"kubernetes.io/projected/2aee299e-6010-491d-bff8-81b4ef2a52c5-kube-api-access-mxxs9\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-2f94z\" (UID: \"2aee299e-6010-491d-bff8-81b4ef2a52c5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2f94z" Dec 01 09:11:09 crc kubenswrapper[4873]: I1201 09:11:09.509339 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2aee299e-6010-491d-bff8-81b4ef2a52c5-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-2f94z\" (UID: \"2aee299e-6010-491d-bff8-81b4ef2a52c5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2f94z" Dec 01 09:11:09 crc kubenswrapper[4873]: I1201 09:11:09.511106 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2aee299e-6010-491d-bff8-81b4ef2a52c5-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-2f94z\" (UID: \"2aee299e-6010-491d-bff8-81b4ef2a52c5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2f94z" Dec 01 09:11:09 crc kubenswrapper[4873]: I1201 09:11:09.526908 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxxs9\" (UniqueName: \"kubernetes.io/projected/2aee299e-6010-491d-bff8-81b4ef2a52c5-kube-api-access-mxxs9\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-2f94z\" (UID: \"2aee299e-6010-491d-bff8-81b4ef2a52c5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2f94z" Dec 01 09:11:09 crc kubenswrapper[4873]: I1201 09:11:09.568503 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2f94z" Dec 01 09:11:10 crc kubenswrapper[4873]: I1201 09:11:10.139416 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2f94z"] Dec 01 09:11:10 crc kubenswrapper[4873]: I1201 09:11:10.172040 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2f94z" event={"ID":"2aee299e-6010-491d-bff8-81b4ef2a52c5","Type":"ContainerStarted","Data":"24310a98045d97a1d5a6d812bbd44cf7fc0165ebce845f6d84e7ef862a761a31"} Dec 01 09:11:11 crc kubenswrapper[4873]: I1201 09:11:11.187550 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2f94z" event={"ID":"2aee299e-6010-491d-bff8-81b4ef2a52c5","Type":"ContainerStarted","Data":"f62011a713228303254b933fa9fb59f221386364ba14686a1da3aa9692dd3eda"} Dec 01 09:11:11 crc kubenswrapper[4873]: I1201 09:11:11.207742 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2f94z" podStartSLOduration=1.7503185650000002 podStartE2EDuration="2.207711523s" podCreationTimestamp="2025-12-01 09:11:09 +0000 UTC" firstStartedPulling="2025-12-01 09:11:10.151879702 +0000 UTC m=+1846.053988241" lastFinishedPulling="2025-12-01 09:11:10.60927265 +0000 UTC m=+1846.511381199" observedRunningTime="2025-12-01 09:11:11.204956275 +0000 UTC m=+1847.107064814" watchObservedRunningTime="2025-12-01 09:11:11.207711523 +0000 UTC m=+1847.109820062" Dec 01 09:11:20 crc kubenswrapper[4873]: I1201 09:11:20.285588 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2f94z" event={"ID":"2aee299e-6010-491d-bff8-81b4ef2a52c5","Type":"ContainerDied","Data":"f62011a713228303254b933fa9fb59f221386364ba14686a1da3aa9692dd3eda"} Dec 01 09:11:20 crc kubenswrapper[4873]: I1201 09:11:20.285488 4873 generic.go:334] "Generic (PLEG): container finished" podID="2aee299e-6010-491d-bff8-81b4ef2a52c5" containerID="f62011a713228303254b933fa9fb59f221386364ba14686a1da3aa9692dd3eda" exitCode=0 Dec 01 09:11:21 crc kubenswrapper[4873]: I1201 09:11:21.729271 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2f94z" Dec 01 09:11:21 crc kubenswrapper[4873]: I1201 09:11:21.880818 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2aee299e-6010-491d-bff8-81b4ef2a52c5-ssh-key\") pod \"2aee299e-6010-491d-bff8-81b4ef2a52c5\" (UID: \"2aee299e-6010-491d-bff8-81b4ef2a52c5\") " Dec 01 09:11:21 crc kubenswrapper[4873]: I1201 09:11:21.880971 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mxxs9\" (UniqueName: \"kubernetes.io/projected/2aee299e-6010-491d-bff8-81b4ef2a52c5-kube-api-access-mxxs9\") pod \"2aee299e-6010-491d-bff8-81b4ef2a52c5\" (UID: \"2aee299e-6010-491d-bff8-81b4ef2a52c5\") " Dec 01 09:11:21 crc kubenswrapper[4873]: I1201 09:11:21.881104 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2aee299e-6010-491d-bff8-81b4ef2a52c5-inventory\") pod \"2aee299e-6010-491d-bff8-81b4ef2a52c5\" (UID: \"2aee299e-6010-491d-bff8-81b4ef2a52c5\") " Dec 01 09:11:21 crc kubenswrapper[4873]: I1201 09:11:21.890763 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2aee299e-6010-491d-bff8-81b4ef2a52c5-kube-api-access-mxxs9" (OuterVolumeSpecName: "kube-api-access-mxxs9") pod "2aee299e-6010-491d-bff8-81b4ef2a52c5" (UID: "2aee299e-6010-491d-bff8-81b4ef2a52c5"). InnerVolumeSpecName "kube-api-access-mxxs9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:11:21 crc kubenswrapper[4873]: I1201 09:11:21.910928 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aee299e-6010-491d-bff8-81b4ef2a52c5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2aee299e-6010-491d-bff8-81b4ef2a52c5" (UID: "2aee299e-6010-491d-bff8-81b4ef2a52c5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:11:21 crc kubenswrapper[4873]: I1201 09:11:21.913388 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aee299e-6010-491d-bff8-81b4ef2a52c5-inventory" (OuterVolumeSpecName: "inventory") pod "2aee299e-6010-491d-bff8-81b4ef2a52c5" (UID: "2aee299e-6010-491d-bff8-81b4ef2a52c5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:11:21 crc kubenswrapper[4873]: I1201 09:11:21.983933 4873 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2aee299e-6010-491d-bff8-81b4ef2a52c5-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 09:11:21 crc kubenswrapper[4873]: I1201 09:11:21.983976 4873 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2aee299e-6010-491d-bff8-81b4ef2a52c5-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 09:11:21 crc kubenswrapper[4873]: I1201 09:11:21.983987 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mxxs9\" (UniqueName: \"kubernetes.io/projected/2aee299e-6010-491d-bff8-81b4ef2a52c5-kube-api-access-mxxs9\") on node \"crc\" DevicePath \"\"" Dec 01 09:11:22 crc kubenswrapper[4873]: I1201 09:11:22.308201 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2f94z" event={"ID":"2aee299e-6010-491d-bff8-81b4ef2a52c5","Type":"ContainerDied","Data":"24310a98045d97a1d5a6d812bbd44cf7fc0165ebce845f6d84e7ef862a761a31"} Dec 01 09:11:22 crc kubenswrapper[4873]: I1201 09:11:22.308282 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="24310a98045d97a1d5a6d812bbd44cf7fc0165ebce845f6d84e7ef862a761a31" Dec 01 09:11:22 crc kubenswrapper[4873]: I1201 09:11:22.308280 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2f94z" Dec 01 09:11:32 crc kubenswrapper[4873]: I1201 09:11:32.051729 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-n2m5g"] Dec 01 09:11:32 crc kubenswrapper[4873]: I1201 09:11:32.070911 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-n2m5g"] Dec 01 09:11:32 crc kubenswrapper[4873]: I1201 09:11:32.446884 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53150db8-4612-45c3-a892-8d6619a82a2f" path="/var/lib/kubelet/pods/53150db8-4612-45c3-a892-8d6619a82a2f/volumes" Dec 01 09:11:52 crc kubenswrapper[4873]: I1201 09:11:52.037989 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-jmbjv"] Dec 01 09:11:52 crc kubenswrapper[4873]: I1201 09:11:52.047996 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-2n4lc"] Dec 01 09:11:52 crc kubenswrapper[4873]: I1201 09:11:52.060177 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-jmbjv"] Dec 01 09:11:52 crc kubenswrapper[4873]: I1201 09:11:52.070340 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-2n4lc"] Dec 01 09:11:52 crc kubenswrapper[4873]: I1201 09:11:52.442976 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48fbd58f-975d-45df-bcf1-7054a4b8abe3" path="/var/lib/kubelet/pods/48fbd58f-975d-45df-bcf1-7054a4b8abe3/volumes" Dec 01 09:11:52 crc kubenswrapper[4873]: I1201 09:11:52.443979 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d" path="/var/lib/kubelet/pods/9e2eb8ff-2d75-4ba8-b4b2-c0bd6723fb9d/volumes" Dec 01 09:11:56 crc kubenswrapper[4873]: I1201 09:11:56.897268 4873 scope.go:117] "RemoveContainer" containerID="4ed7f8b8d815f149f614aad11fde49b28ef841b0121fd81a73b761c17fc3e112" Dec 01 09:11:56 crc kubenswrapper[4873]: I1201 09:11:56.949574 4873 scope.go:117] "RemoveContainer" containerID="523238858a5e7d371b2e3b8d87e743130e7c243cfc9d9ebddaf8d98fb42483a2" Dec 01 09:11:56 crc kubenswrapper[4873]: I1201 09:11:56.972711 4873 scope.go:117] "RemoveContainer" containerID="ff034402aa98a628762b457748842a51a9fcaee7130f2a41fcd880a4ea5d9545" Dec 01 09:11:57 crc kubenswrapper[4873]: I1201 09:11:57.040123 4873 scope.go:117] "RemoveContainer" containerID="29199103bf26d5451c7186e11bf49db090317374b7bcc88aceb00d4ec66e563e" Dec 01 09:11:57 crc kubenswrapper[4873]: I1201 09:11:57.077249 4873 scope.go:117] "RemoveContainer" containerID="cbc26a4e9818a65332ccad8daf87f3c64050554ff7c97689265a4d9c5971f653" Dec 01 09:11:57 crc kubenswrapper[4873]: I1201 09:11:57.114745 4873 scope.go:117] "RemoveContainer" containerID="6d96f4b4c03294ca19a5cac036c6ad6e9bdc9fbff00ac03ebb291a0f6853705f" Dec 01 09:11:57 crc kubenswrapper[4873]: I1201 09:11:57.158494 4873 scope.go:117] "RemoveContainer" containerID="5642159ae93809131ed84df354644b68602e4203c88bc6f2286dc6474f8a2232" Dec 01 09:11:57 crc kubenswrapper[4873]: I1201 09:11:57.204514 4873 scope.go:117] "RemoveContainer" containerID="5ba249e9ca335fc03a2d70f8e3a190f77197791d48ef944e6c4d4574c5c5276e" Dec 01 09:11:57 crc kubenswrapper[4873]: I1201 09:11:57.226166 4873 scope.go:117] "RemoveContainer" containerID="928b07338f76deaadc48dcb860deff8b6ee9693b1b0dd0ab0d01c63540831087" Dec 01 09:12:37 crc kubenswrapper[4873]: I1201 09:12:37.059392 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-h4n9r"] Dec 01 09:12:37 crc kubenswrapper[4873]: I1201 09:12:37.071415 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-h4n9r"] Dec 01 09:12:38 crc kubenswrapper[4873]: I1201 09:12:38.441750 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a911632-e929-4993-9f93-f88d6e6c8993" path="/var/lib/kubelet/pods/0a911632-e929-4993-9f93-f88d6e6c8993/volumes" Dec 01 09:12:57 crc kubenswrapper[4873]: I1201 09:12:57.412752 4873 scope.go:117] "RemoveContainer" containerID="410b00412bb25258ee0029e32b927f1d745f3d5ab7fb21f04a4bd75930426a37" Dec 01 09:12:57 crc kubenswrapper[4873]: I1201 09:12:57.479583 4873 scope.go:117] "RemoveContainer" containerID="ef1a432d2e318e827260ee3123cccc8d2c486baac6fbb6f3aaeef3b139f9150f" Dec 01 09:12:57 crc kubenswrapper[4873]: I1201 09:12:57.503413 4873 scope.go:117] "RemoveContainer" containerID="5657f2f0bad208842ecb40eb62da42f41bf72d7628670666729dfa7ab366a642" Dec 01 09:12:57 crc kubenswrapper[4873]: I1201 09:12:57.525625 4873 scope.go:117] "RemoveContainer" containerID="45dd4a2e9ef3e533528e525385402cc0c43b478549e7199f9e32a5402432ef0c" Dec 01 09:13:01 crc kubenswrapper[4873]: I1201 09:13:01.059920 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:13:01 crc kubenswrapper[4873]: I1201 09:13:01.060591 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:13:31 crc kubenswrapper[4873]: I1201 09:13:31.059460 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:13:31 crc kubenswrapper[4873]: I1201 09:13:31.060336 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:14:01 crc kubenswrapper[4873]: I1201 09:14:01.059182 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:14:01 crc kubenswrapper[4873]: I1201 09:14:01.061008 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:14:01 crc kubenswrapper[4873]: I1201 09:14:01.061180 4873 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" Dec 01 09:14:01 crc kubenswrapper[4873]: I1201 09:14:01.062421 4873 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"281f1071cdd39ef7daaacd7ebb1e90ad778c28e93bebe102e714733c9cf3c602"} pod="openshift-machine-config-operator/machine-config-daemon-scwpp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 09:14:01 crc kubenswrapper[4873]: I1201 09:14:01.062585 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" containerID="cri-o://281f1071cdd39ef7daaacd7ebb1e90ad778c28e93bebe102e714733c9cf3c602" gracePeriod=600 Dec 01 09:14:01 crc kubenswrapper[4873]: I1201 09:14:01.443595 4873 generic.go:334] "Generic (PLEG): container finished" podID="fef7b114-0e07-402d-a37b-315c36011f4b" containerID="281f1071cdd39ef7daaacd7ebb1e90ad778c28e93bebe102e714733c9cf3c602" exitCode=0 Dec 01 09:14:01 crc kubenswrapper[4873]: I1201 09:14:01.443663 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" event={"ID":"fef7b114-0e07-402d-a37b-315c36011f4b","Type":"ContainerDied","Data":"281f1071cdd39ef7daaacd7ebb1e90ad778c28e93bebe102e714733c9cf3c602"} Dec 01 09:14:01 crc kubenswrapper[4873]: I1201 09:14:01.444188 4873 scope.go:117] "RemoveContainer" containerID="dd1fd5c773938eaca06d208e8759b6376ff65c7cad3dc995d4b7c31e75f098dc" Dec 01 09:14:02 crc kubenswrapper[4873]: I1201 09:14:02.456680 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" event={"ID":"fef7b114-0e07-402d-a37b-315c36011f4b","Type":"ContainerStarted","Data":"1a1db8f8327e8cf6dbb94b57ea898b250339bb198ae1ba595cb1d79805fa0db1"} Dec 01 09:15:00 crc kubenswrapper[4873]: I1201 09:15:00.157911 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409675-ksw4s"] Dec 01 09:15:00 crc kubenswrapper[4873]: E1201 09:15:00.159572 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2aee299e-6010-491d-bff8-81b4ef2a52c5" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 01 09:15:00 crc kubenswrapper[4873]: I1201 09:15:00.159600 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="2aee299e-6010-491d-bff8-81b4ef2a52c5" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 01 09:15:00 crc kubenswrapper[4873]: I1201 09:15:00.159912 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="2aee299e-6010-491d-bff8-81b4ef2a52c5" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 01 09:15:00 crc kubenswrapper[4873]: I1201 09:15:00.161184 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-ksw4s" Dec 01 09:15:00 crc kubenswrapper[4873]: I1201 09:15:00.164108 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 09:15:00 crc kubenswrapper[4873]: I1201 09:15:00.166082 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 09:15:00 crc kubenswrapper[4873]: I1201 09:15:00.171644 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409675-ksw4s"] Dec 01 09:15:00 crc kubenswrapper[4873]: I1201 09:15:00.228917 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b3e6f863-4d43-4533-91bb-4ac8ae51a645-config-volume\") pod \"collect-profiles-29409675-ksw4s\" (UID: \"b3e6f863-4d43-4533-91bb-4ac8ae51a645\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-ksw4s" Dec 01 09:15:00 crc kubenswrapper[4873]: I1201 09:15:00.228989 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b3e6f863-4d43-4533-91bb-4ac8ae51a645-secret-volume\") pod \"collect-profiles-29409675-ksw4s\" (UID: \"b3e6f863-4d43-4533-91bb-4ac8ae51a645\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-ksw4s" Dec 01 09:15:00 crc kubenswrapper[4873]: I1201 09:15:00.229549 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xwl8\" (UniqueName: \"kubernetes.io/projected/b3e6f863-4d43-4533-91bb-4ac8ae51a645-kube-api-access-6xwl8\") pod \"collect-profiles-29409675-ksw4s\" (UID: \"b3e6f863-4d43-4533-91bb-4ac8ae51a645\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-ksw4s" Dec 01 09:15:00 crc kubenswrapper[4873]: I1201 09:15:00.331349 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b3e6f863-4d43-4533-91bb-4ac8ae51a645-config-volume\") pod \"collect-profiles-29409675-ksw4s\" (UID: \"b3e6f863-4d43-4533-91bb-4ac8ae51a645\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-ksw4s" Dec 01 09:15:00 crc kubenswrapper[4873]: I1201 09:15:00.331407 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b3e6f863-4d43-4533-91bb-4ac8ae51a645-secret-volume\") pod \"collect-profiles-29409675-ksw4s\" (UID: \"b3e6f863-4d43-4533-91bb-4ac8ae51a645\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-ksw4s" Dec 01 09:15:00 crc kubenswrapper[4873]: I1201 09:15:00.331540 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xwl8\" (UniqueName: \"kubernetes.io/projected/b3e6f863-4d43-4533-91bb-4ac8ae51a645-kube-api-access-6xwl8\") pod \"collect-profiles-29409675-ksw4s\" (UID: \"b3e6f863-4d43-4533-91bb-4ac8ae51a645\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-ksw4s" Dec 01 09:15:00 crc kubenswrapper[4873]: I1201 09:15:00.332824 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b3e6f863-4d43-4533-91bb-4ac8ae51a645-config-volume\") pod \"collect-profiles-29409675-ksw4s\" (UID: \"b3e6f863-4d43-4533-91bb-4ac8ae51a645\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-ksw4s" Dec 01 09:15:00 crc kubenswrapper[4873]: I1201 09:15:00.339606 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b3e6f863-4d43-4533-91bb-4ac8ae51a645-secret-volume\") pod \"collect-profiles-29409675-ksw4s\" (UID: \"b3e6f863-4d43-4533-91bb-4ac8ae51a645\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-ksw4s" Dec 01 09:15:00 crc kubenswrapper[4873]: I1201 09:15:00.349841 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xwl8\" (UniqueName: \"kubernetes.io/projected/b3e6f863-4d43-4533-91bb-4ac8ae51a645-kube-api-access-6xwl8\") pod \"collect-profiles-29409675-ksw4s\" (UID: \"b3e6f863-4d43-4533-91bb-4ac8ae51a645\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-ksw4s" Dec 01 09:15:00 crc kubenswrapper[4873]: I1201 09:15:00.496392 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-ksw4s" Dec 01 09:15:01 crc kubenswrapper[4873]: I1201 09:15:01.012524 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409675-ksw4s"] Dec 01 09:15:01 crc kubenswrapper[4873]: I1201 09:15:01.532917 4873 generic.go:334] "Generic (PLEG): container finished" podID="b3e6f863-4d43-4533-91bb-4ac8ae51a645" containerID="475e5604dc0257b8aa1e561a9897be97cca92ba1ce70ba0c1be43911f6524a46" exitCode=0 Dec 01 09:15:01 crc kubenswrapper[4873]: I1201 09:15:01.533123 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-ksw4s" event={"ID":"b3e6f863-4d43-4533-91bb-4ac8ae51a645","Type":"ContainerDied","Data":"475e5604dc0257b8aa1e561a9897be97cca92ba1ce70ba0c1be43911f6524a46"} Dec 01 09:15:01 crc kubenswrapper[4873]: I1201 09:15:01.533422 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-ksw4s" event={"ID":"b3e6f863-4d43-4533-91bb-4ac8ae51a645","Type":"ContainerStarted","Data":"91ee5cfecb70904c1cebccc1d7beb346180caa33a7203bbb632e5e549d3c8452"} Dec 01 09:15:02 crc kubenswrapper[4873]: I1201 09:15:02.871849 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-ksw4s" Dec 01 09:15:03 crc kubenswrapper[4873]: I1201 09:15:03.008191 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b3e6f863-4d43-4533-91bb-4ac8ae51a645-secret-volume\") pod \"b3e6f863-4d43-4533-91bb-4ac8ae51a645\" (UID: \"b3e6f863-4d43-4533-91bb-4ac8ae51a645\") " Dec 01 09:15:03 crc kubenswrapper[4873]: I1201 09:15:03.008837 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6xwl8\" (UniqueName: \"kubernetes.io/projected/b3e6f863-4d43-4533-91bb-4ac8ae51a645-kube-api-access-6xwl8\") pod \"b3e6f863-4d43-4533-91bb-4ac8ae51a645\" (UID: \"b3e6f863-4d43-4533-91bb-4ac8ae51a645\") " Dec 01 09:15:03 crc kubenswrapper[4873]: I1201 09:15:03.008890 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b3e6f863-4d43-4533-91bb-4ac8ae51a645-config-volume\") pod \"b3e6f863-4d43-4533-91bb-4ac8ae51a645\" (UID: \"b3e6f863-4d43-4533-91bb-4ac8ae51a645\") " Dec 01 09:15:03 crc kubenswrapper[4873]: I1201 09:15:03.010187 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3e6f863-4d43-4533-91bb-4ac8ae51a645-config-volume" (OuterVolumeSpecName: "config-volume") pod "b3e6f863-4d43-4533-91bb-4ac8ae51a645" (UID: "b3e6f863-4d43-4533-91bb-4ac8ae51a645"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:15:03 crc kubenswrapper[4873]: I1201 09:15:03.022349 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3e6f863-4d43-4533-91bb-4ac8ae51a645-kube-api-access-6xwl8" (OuterVolumeSpecName: "kube-api-access-6xwl8") pod "b3e6f863-4d43-4533-91bb-4ac8ae51a645" (UID: "b3e6f863-4d43-4533-91bb-4ac8ae51a645"). InnerVolumeSpecName "kube-api-access-6xwl8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:15:03 crc kubenswrapper[4873]: I1201 09:15:03.022593 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3e6f863-4d43-4533-91bb-4ac8ae51a645-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b3e6f863-4d43-4533-91bb-4ac8ae51a645" (UID: "b3e6f863-4d43-4533-91bb-4ac8ae51a645"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:15:03 crc kubenswrapper[4873]: I1201 09:15:03.111180 4873 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b3e6f863-4d43-4533-91bb-4ac8ae51a645-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 09:15:03 crc kubenswrapper[4873]: I1201 09:15:03.111235 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6xwl8\" (UniqueName: \"kubernetes.io/projected/b3e6f863-4d43-4533-91bb-4ac8ae51a645-kube-api-access-6xwl8\") on node \"crc\" DevicePath \"\"" Dec 01 09:15:03 crc kubenswrapper[4873]: I1201 09:15:03.111251 4873 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b3e6f863-4d43-4533-91bb-4ac8ae51a645-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 09:15:03 crc kubenswrapper[4873]: I1201 09:15:03.553979 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-ksw4s" event={"ID":"b3e6f863-4d43-4533-91bb-4ac8ae51a645","Type":"ContainerDied","Data":"91ee5cfecb70904c1cebccc1d7beb346180caa33a7203bbb632e5e549d3c8452"} Dec 01 09:15:03 crc kubenswrapper[4873]: I1201 09:15:03.554056 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-ksw4s" Dec 01 09:15:03 crc kubenswrapper[4873]: I1201 09:15:03.554071 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="91ee5cfecb70904c1cebccc1d7beb346180caa33a7203bbb632e5e549d3c8452" Dec 01 09:15:03 crc kubenswrapper[4873]: I1201 09:15:03.958308 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409630-lnw2h"] Dec 01 09:15:03 crc kubenswrapper[4873]: I1201 09:15:03.966810 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409630-lnw2h"] Dec 01 09:15:04 crc kubenswrapper[4873]: I1201 09:15:04.442522 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83688ccb-ace0-4b0a-b584-e650671f5686" path="/var/lib/kubelet/pods/83688ccb-ace0-4b0a-b584-e650671f5686/volumes" Dec 01 09:15:50 crc kubenswrapper[4873]: I1201 09:15:50.692523 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ffspk"] Dec 01 09:15:50 crc kubenswrapper[4873]: I1201 09:15:50.712527 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-j2z8p"] Dec 01 09:15:50 crc kubenswrapper[4873]: I1201 09:15:50.722564 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tsk8x"] Dec 01 09:15:50 crc kubenswrapper[4873]: I1201 09:15:50.735131 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-sgwcl"] Dec 01 09:15:50 crc kubenswrapper[4873]: I1201 09:15:50.745493 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-f9cnt"] Dec 01 09:15:50 crc kubenswrapper[4873]: I1201 09:15:50.755851 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ffspk"] Dec 01 09:15:50 crc kubenswrapper[4873]: I1201 09:15:50.768224 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2f94z"] Dec 01 09:15:50 crc kubenswrapper[4873]: I1201 09:15:50.778610 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-j2z8p"] Dec 01 09:15:50 crc kubenswrapper[4873]: I1201 09:15:50.789663 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-f9cnt"] Dec 01 09:15:50 crc kubenswrapper[4873]: I1201 09:15:50.798981 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tsk8x"] Dec 01 09:15:50 crc kubenswrapper[4873]: I1201 09:15:50.806169 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-bbwph"] Dec 01 09:15:50 crc kubenswrapper[4873]: I1201 09:15:50.814338 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-5x9zj"] Dec 01 09:15:50 crc kubenswrapper[4873]: I1201 09:15:50.823485 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-sgwcl"] Dec 01 09:15:50 crc kubenswrapper[4873]: I1201 09:15:50.829956 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dhxtp"] Dec 01 09:15:50 crc kubenswrapper[4873]: I1201 09:15:50.840172 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2f94z"] Dec 01 09:15:50 crc kubenswrapper[4873]: I1201 09:15:50.844978 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wfjd2"] Dec 01 09:15:50 crc kubenswrapper[4873]: I1201 09:15:50.852205 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-5x9zj"] Dec 01 09:15:50 crc kubenswrapper[4873]: I1201 09:15:50.859582 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dhxtp"] Dec 01 09:15:50 crc kubenswrapper[4873]: I1201 09:15:50.866695 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-bbwph"] Dec 01 09:15:50 crc kubenswrapper[4873]: I1201 09:15:50.872987 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wfjd2"] Dec 01 09:15:52 crc kubenswrapper[4873]: I1201 09:15:52.456093 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1402ba59-201c-4b03-8c0e-478c9d560ef5" path="/var/lib/kubelet/pods/1402ba59-201c-4b03-8c0e-478c9d560ef5/volumes" Dec 01 09:15:52 crc kubenswrapper[4873]: I1201 09:15:52.457578 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1" path="/var/lib/kubelet/pods/2a411d16-2d81-4ca8-b99c-92ad0b2eb9e1/volumes" Dec 01 09:15:52 crc kubenswrapper[4873]: I1201 09:15:52.458507 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2aee299e-6010-491d-bff8-81b4ef2a52c5" path="/var/lib/kubelet/pods/2aee299e-6010-491d-bff8-81b4ef2a52c5/volumes" Dec 01 09:15:52 crc kubenswrapper[4873]: I1201 09:15:52.459502 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="489b7faf-eb94-459f-b279-7b5cedf53db9" path="/var/lib/kubelet/pods/489b7faf-eb94-459f-b279-7b5cedf53db9/volumes" Dec 01 09:15:52 crc kubenswrapper[4873]: I1201 09:15:52.461302 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="829ed711-4da5-47ed-80c2-4af8aff6e84a" path="/var/lib/kubelet/pods/829ed711-4da5-47ed-80c2-4af8aff6e84a/volumes" Dec 01 09:15:52 crc kubenswrapper[4873]: I1201 09:15:52.462142 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e0f903d-9fb9-4b2b-9194-a9316d8b796b" path="/var/lib/kubelet/pods/8e0f903d-9fb9-4b2b-9194-a9316d8b796b/volumes" Dec 01 09:15:52 crc kubenswrapper[4873]: I1201 09:15:52.463185 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f853021-83c9-47c0-bf56-7374ffc662cf" path="/var/lib/kubelet/pods/9f853021-83c9-47c0-bf56-7374ffc662cf/volumes" Dec 01 09:15:52 crc kubenswrapper[4873]: I1201 09:15:52.465516 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0" path="/var/lib/kubelet/pods/a0b9ce0e-b2fd-4e34-9be2-ff3f9a54abc0/volumes" Dec 01 09:15:52 crc kubenswrapper[4873]: I1201 09:15:52.466479 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd" path="/var/lib/kubelet/pods/ba3d9d8a-e00b-467f-b718-d5fcfeea1fdd/volumes" Dec 01 09:15:52 crc kubenswrapper[4873]: I1201 09:15:52.467461 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c34d2445-0767-413e-a35e-61258e34e344" path="/var/lib/kubelet/pods/c34d2445-0767-413e-a35e-61258e34e344/volumes" Dec 01 09:15:56 crc kubenswrapper[4873]: I1201 09:15:56.598126 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g5ghr"] Dec 01 09:15:56 crc kubenswrapper[4873]: E1201 09:15:56.599200 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3e6f863-4d43-4533-91bb-4ac8ae51a645" containerName="collect-profiles" Dec 01 09:15:56 crc kubenswrapper[4873]: I1201 09:15:56.599220 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3e6f863-4d43-4533-91bb-4ac8ae51a645" containerName="collect-profiles" Dec 01 09:15:56 crc kubenswrapper[4873]: I1201 09:15:56.599457 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3e6f863-4d43-4533-91bb-4ac8ae51a645" containerName="collect-profiles" Dec 01 09:15:56 crc kubenswrapper[4873]: I1201 09:15:56.600307 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g5ghr" Dec 01 09:15:56 crc kubenswrapper[4873]: I1201 09:15:56.604950 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 09:15:56 crc kubenswrapper[4873]: I1201 09:15:56.605388 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9ghqw" Dec 01 09:15:56 crc kubenswrapper[4873]: I1201 09:15:56.605413 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 09:15:56 crc kubenswrapper[4873]: I1201 09:15:56.605739 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 09:15:56 crc kubenswrapper[4873]: I1201 09:15:56.605918 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 01 09:15:56 crc kubenswrapper[4873]: I1201 09:15:56.622706 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g5ghr"] Dec 01 09:15:56 crc kubenswrapper[4873]: I1201 09:15:56.713345 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/47baf114-fcb4-46ba-ac64-53f969ebcbbc-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g5ghr\" (UID: \"47baf114-fcb4-46ba-ac64-53f969ebcbbc\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g5ghr" Dec 01 09:15:56 crc kubenswrapper[4873]: I1201 09:15:56.713449 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjn7j\" (UniqueName: \"kubernetes.io/projected/47baf114-fcb4-46ba-ac64-53f969ebcbbc-kube-api-access-vjn7j\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g5ghr\" (UID: \"47baf114-fcb4-46ba-ac64-53f969ebcbbc\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g5ghr" Dec 01 09:15:56 crc kubenswrapper[4873]: I1201 09:15:56.713562 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47baf114-fcb4-46ba-ac64-53f969ebcbbc-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g5ghr\" (UID: \"47baf114-fcb4-46ba-ac64-53f969ebcbbc\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g5ghr" Dec 01 09:15:56 crc kubenswrapper[4873]: I1201 09:15:56.713626 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/47baf114-fcb4-46ba-ac64-53f969ebcbbc-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g5ghr\" (UID: \"47baf114-fcb4-46ba-ac64-53f969ebcbbc\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g5ghr" Dec 01 09:15:56 crc kubenswrapper[4873]: I1201 09:15:56.713658 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/47baf114-fcb4-46ba-ac64-53f969ebcbbc-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g5ghr\" (UID: \"47baf114-fcb4-46ba-ac64-53f969ebcbbc\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g5ghr" Dec 01 09:15:56 crc kubenswrapper[4873]: I1201 09:15:56.815635 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/47baf114-fcb4-46ba-ac64-53f969ebcbbc-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g5ghr\" (UID: \"47baf114-fcb4-46ba-ac64-53f969ebcbbc\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g5ghr" Dec 01 09:15:56 crc kubenswrapper[4873]: I1201 09:15:56.815696 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/47baf114-fcb4-46ba-ac64-53f969ebcbbc-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g5ghr\" (UID: \"47baf114-fcb4-46ba-ac64-53f969ebcbbc\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g5ghr" Dec 01 09:15:56 crc kubenswrapper[4873]: I1201 09:15:56.815751 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/47baf114-fcb4-46ba-ac64-53f969ebcbbc-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g5ghr\" (UID: \"47baf114-fcb4-46ba-ac64-53f969ebcbbc\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g5ghr" Dec 01 09:15:56 crc kubenswrapper[4873]: I1201 09:15:56.815790 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjn7j\" (UniqueName: \"kubernetes.io/projected/47baf114-fcb4-46ba-ac64-53f969ebcbbc-kube-api-access-vjn7j\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g5ghr\" (UID: \"47baf114-fcb4-46ba-ac64-53f969ebcbbc\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g5ghr" Dec 01 09:15:56 crc kubenswrapper[4873]: I1201 09:15:56.815877 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47baf114-fcb4-46ba-ac64-53f969ebcbbc-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g5ghr\" (UID: \"47baf114-fcb4-46ba-ac64-53f969ebcbbc\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g5ghr" Dec 01 09:15:56 crc kubenswrapper[4873]: I1201 09:15:56.823655 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/47baf114-fcb4-46ba-ac64-53f969ebcbbc-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g5ghr\" (UID: \"47baf114-fcb4-46ba-ac64-53f969ebcbbc\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g5ghr" Dec 01 09:15:56 crc kubenswrapper[4873]: I1201 09:15:56.823667 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/47baf114-fcb4-46ba-ac64-53f969ebcbbc-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g5ghr\" (UID: \"47baf114-fcb4-46ba-ac64-53f969ebcbbc\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g5ghr" Dec 01 09:15:56 crc kubenswrapper[4873]: I1201 09:15:56.825095 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47baf114-fcb4-46ba-ac64-53f969ebcbbc-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g5ghr\" (UID: \"47baf114-fcb4-46ba-ac64-53f969ebcbbc\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g5ghr" Dec 01 09:15:56 crc kubenswrapper[4873]: I1201 09:15:56.830327 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/47baf114-fcb4-46ba-ac64-53f969ebcbbc-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g5ghr\" (UID: \"47baf114-fcb4-46ba-ac64-53f969ebcbbc\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g5ghr" Dec 01 09:15:56 crc kubenswrapper[4873]: I1201 09:15:56.839554 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjn7j\" (UniqueName: \"kubernetes.io/projected/47baf114-fcb4-46ba-ac64-53f969ebcbbc-kube-api-access-vjn7j\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g5ghr\" (UID: \"47baf114-fcb4-46ba-ac64-53f969ebcbbc\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g5ghr" Dec 01 09:15:56 crc kubenswrapper[4873]: I1201 09:15:56.933593 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g5ghr" Dec 01 09:15:57 crc kubenswrapper[4873]: I1201 09:15:57.671106 4873 scope.go:117] "RemoveContainer" containerID="38b8fc7cc310a3ec72fe0ed1ba4dbfa51153edeb1fc2eb5f97423a83c30f7bc2" Dec 01 09:15:57 crc kubenswrapper[4873]: I1201 09:15:57.714850 4873 scope.go:117] "RemoveContainer" containerID="4f7139adfaa2cc1c1c281536f7a1d5644b92edde4af7aa21aa62b1a7d98fbfc7" Dec 01 09:15:57 crc kubenswrapper[4873]: I1201 09:15:57.800061 4873 scope.go:117] "RemoveContainer" containerID="0b17315b3b408f1a71b69372e87fd41c476a1d6148d2474063bce7ce3f666609" Dec 01 09:15:57 crc kubenswrapper[4873]: I1201 09:15:57.865282 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g5ghr"] Dec 01 09:15:57 crc kubenswrapper[4873]: I1201 09:15:57.901231 4873 scope.go:117] "RemoveContainer" containerID="dfac8b2123466a36a436429e744cce54ec673ca0e1a5b3c80ea5a8ace985a68b" Dec 01 09:15:57 crc kubenswrapper[4873]: I1201 09:15:57.912001 4873 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 09:15:57 crc kubenswrapper[4873]: I1201 09:15:57.981030 4873 scope.go:117] "RemoveContainer" containerID="7a95a64429616277f401a800e029c011e2a05c0a881b88e6a0699f5863f84145" Dec 01 09:15:58 crc kubenswrapper[4873]: I1201 09:15:58.041760 4873 scope.go:117] "RemoveContainer" containerID="754a7b2516f6e400aeaac04d1e0d672b23e9e0fe8b2c778716b9489b5138574e" Dec 01 09:15:58 crc kubenswrapper[4873]: I1201 09:15:58.113906 4873 scope.go:117] "RemoveContainer" containerID="9b0a6483fa36f088701b1a9ba0005c29790f19efacef6f133fba1772329d04ca" Dec 01 09:15:58 crc kubenswrapper[4873]: I1201 09:15:58.151866 4873 scope.go:117] "RemoveContainer" containerID="6a4fdb6abebcf86cf7aaa93b96ee5520f76d39994e852cbb8ec69897ce253fa3" Dec 01 09:15:58 crc kubenswrapper[4873]: I1201 09:15:58.398696 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g5ghr" event={"ID":"47baf114-fcb4-46ba-ac64-53f969ebcbbc","Type":"ContainerStarted","Data":"15301b79d34344e295d52c7d57c32e26634994e458b3e6153f1c9e252483b7f8"} Dec 01 09:15:59 crc kubenswrapper[4873]: I1201 09:15:59.411313 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g5ghr" event={"ID":"47baf114-fcb4-46ba-ac64-53f969ebcbbc","Type":"ContainerStarted","Data":"7265d5af3e6a7fcdd432826702179f6c27eb320558d1c2d331e4d48a9211fc01"} Dec 01 09:15:59 crc kubenswrapper[4873]: I1201 09:15:59.441516 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g5ghr" podStartSLOduration=2.371104442 podStartE2EDuration="3.441489951s" podCreationTimestamp="2025-12-01 09:15:56 +0000 UTC" firstStartedPulling="2025-12-01 09:15:57.911614274 +0000 UTC m=+2133.813722813" lastFinishedPulling="2025-12-01 09:15:58.981999783 +0000 UTC m=+2134.884108322" observedRunningTime="2025-12-01 09:15:59.433256868 +0000 UTC m=+2135.335365407" watchObservedRunningTime="2025-12-01 09:15:59.441489951 +0000 UTC m=+2135.343598490" Dec 01 09:16:01 crc kubenswrapper[4873]: I1201 09:16:01.059243 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:16:01 crc kubenswrapper[4873]: I1201 09:16:01.059685 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:16:11 crc kubenswrapper[4873]: I1201 09:16:11.538153 4873 generic.go:334] "Generic (PLEG): container finished" podID="47baf114-fcb4-46ba-ac64-53f969ebcbbc" containerID="7265d5af3e6a7fcdd432826702179f6c27eb320558d1c2d331e4d48a9211fc01" exitCode=0 Dec 01 09:16:11 crc kubenswrapper[4873]: I1201 09:16:11.538366 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g5ghr" event={"ID":"47baf114-fcb4-46ba-ac64-53f969ebcbbc","Type":"ContainerDied","Data":"7265d5af3e6a7fcdd432826702179f6c27eb320558d1c2d331e4d48a9211fc01"} Dec 01 09:16:12 crc kubenswrapper[4873]: I1201 09:16:12.965363 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g5ghr" Dec 01 09:16:13 crc kubenswrapper[4873]: I1201 09:16:13.109213 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/47baf114-fcb4-46ba-ac64-53f969ebcbbc-ssh-key\") pod \"47baf114-fcb4-46ba-ac64-53f969ebcbbc\" (UID: \"47baf114-fcb4-46ba-ac64-53f969ebcbbc\") " Dec 01 09:16:13 crc kubenswrapper[4873]: I1201 09:16:13.109296 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47baf114-fcb4-46ba-ac64-53f969ebcbbc-repo-setup-combined-ca-bundle\") pod \"47baf114-fcb4-46ba-ac64-53f969ebcbbc\" (UID: \"47baf114-fcb4-46ba-ac64-53f969ebcbbc\") " Dec 01 09:16:13 crc kubenswrapper[4873]: I1201 09:16:13.109322 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/47baf114-fcb4-46ba-ac64-53f969ebcbbc-inventory\") pod \"47baf114-fcb4-46ba-ac64-53f969ebcbbc\" (UID: \"47baf114-fcb4-46ba-ac64-53f969ebcbbc\") " Dec 01 09:16:13 crc kubenswrapper[4873]: I1201 09:16:13.109428 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vjn7j\" (UniqueName: \"kubernetes.io/projected/47baf114-fcb4-46ba-ac64-53f969ebcbbc-kube-api-access-vjn7j\") pod \"47baf114-fcb4-46ba-ac64-53f969ebcbbc\" (UID: \"47baf114-fcb4-46ba-ac64-53f969ebcbbc\") " Dec 01 09:16:13 crc kubenswrapper[4873]: I1201 09:16:13.109518 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/47baf114-fcb4-46ba-ac64-53f969ebcbbc-ceph\") pod \"47baf114-fcb4-46ba-ac64-53f969ebcbbc\" (UID: \"47baf114-fcb4-46ba-ac64-53f969ebcbbc\") " Dec 01 09:16:13 crc kubenswrapper[4873]: I1201 09:16:13.116823 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47baf114-fcb4-46ba-ac64-53f969ebcbbc-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "47baf114-fcb4-46ba-ac64-53f969ebcbbc" (UID: "47baf114-fcb4-46ba-ac64-53f969ebcbbc"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:16:13 crc kubenswrapper[4873]: I1201 09:16:13.120008 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47baf114-fcb4-46ba-ac64-53f969ebcbbc-ceph" (OuterVolumeSpecName: "ceph") pod "47baf114-fcb4-46ba-ac64-53f969ebcbbc" (UID: "47baf114-fcb4-46ba-ac64-53f969ebcbbc"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:16:13 crc kubenswrapper[4873]: I1201 09:16:13.120144 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47baf114-fcb4-46ba-ac64-53f969ebcbbc-kube-api-access-vjn7j" (OuterVolumeSpecName: "kube-api-access-vjn7j") pod "47baf114-fcb4-46ba-ac64-53f969ebcbbc" (UID: "47baf114-fcb4-46ba-ac64-53f969ebcbbc"). InnerVolumeSpecName "kube-api-access-vjn7j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:16:13 crc kubenswrapper[4873]: I1201 09:16:13.139431 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47baf114-fcb4-46ba-ac64-53f969ebcbbc-inventory" (OuterVolumeSpecName: "inventory") pod "47baf114-fcb4-46ba-ac64-53f969ebcbbc" (UID: "47baf114-fcb4-46ba-ac64-53f969ebcbbc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:16:13 crc kubenswrapper[4873]: I1201 09:16:13.143261 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47baf114-fcb4-46ba-ac64-53f969ebcbbc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "47baf114-fcb4-46ba-ac64-53f969ebcbbc" (UID: "47baf114-fcb4-46ba-ac64-53f969ebcbbc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:16:13 crc kubenswrapper[4873]: I1201 09:16:13.213192 4873 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/47baf114-fcb4-46ba-ac64-53f969ebcbbc-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 09:16:13 crc kubenswrapper[4873]: I1201 09:16:13.213242 4873 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47baf114-fcb4-46ba-ac64-53f969ebcbbc-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:16:13 crc kubenswrapper[4873]: I1201 09:16:13.213254 4873 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/47baf114-fcb4-46ba-ac64-53f969ebcbbc-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 09:16:13 crc kubenswrapper[4873]: I1201 09:16:13.213266 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vjn7j\" (UniqueName: \"kubernetes.io/projected/47baf114-fcb4-46ba-ac64-53f969ebcbbc-kube-api-access-vjn7j\") on node \"crc\" DevicePath \"\"" Dec 01 09:16:13 crc kubenswrapper[4873]: I1201 09:16:13.213277 4873 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/47baf114-fcb4-46ba-ac64-53f969ebcbbc-ceph\") on node \"crc\" DevicePath \"\"" Dec 01 09:16:13 crc kubenswrapper[4873]: I1201 09:16:13.563438 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g5ghr" event={"ID":"47baf114-fcb4-46ba-ac64-53f969ebcbbc","Type":"ContainerDied","Data":"15301b79d34344e295d52c7d57c32e26634994e458b3e6153f1c9e252483b7f8"} Dec 01 09:16:13 crc kubenswrapper[4873]: I1201 09:16:13.563501 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="15301b79d34344e295d52c7d57c32e26634994e458b3e6153f1c9e252483b7f8" Dec 01 09:16:13 crc kubenswrapper[4873]: I1201 09:16:13.563567 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g5ghr" Dec 01 09:16:13 crc kubenswrapper[4873]: I1201 09:16:13.700680 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wfccb"] Dec 01 09:16:13 crc kubenswrapper[4873]: E1201 09:16:13.701288 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47baf114-fcb4-46ba-ac64-53f969ebcbbc" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 01 09:16:13 crc kubenswrapper[4873]: I1201 09:16:13.701312 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="47baf114-fcb4-46ba-ac64-53f969ebcbbc" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 01 09:16:13 crc kubenswrapper[4873]: I1201 09:16:13.701527 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="47baf114-fcb4-46ba-ac64-53f969ebcbbc" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 01 09:16:13 crc kubenswrapper[4873]: I1201 09:16:13.702402 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wfccb" Dec 01 09:16:13 crc kubenswrapper[4873]: I1201 09:16:13.708428 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 09:16:13 crc kubenswrapper[4873]: I1201 09:16:13.708547 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wfccb"] Dec 01 09:16:13 crc kubenswrapper[4873]: I1201 09:16:13.708795 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9ghqw" Dec 01 09:16:13 crc kubenswrapper[4873]: I1201 09:16:13.708967 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 09:16:13 crc kubenswrapper[4873]: I1201 09:16:13.710031 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 01 09:16:13 crc kubenswrapper[4873]: I1201 09:16:13.711669 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 09:16:14 crc kubenswrapper[4873]: I1201 09:16:14.087704 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cd2a1185-cebc-4f89-88c2-63cfebe97759-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wfccb\" (UID: \"cd2a1185-cebc-4f89-88c2-63cfebe97759\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wfccb" Dec 01 09:16:14 crc kubenswrapper[4873]: I1201 09:16:14.092385 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd2a1185-cebc-4f89-88c2-63cfebe97759-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wfccb\" (UID: \"cd2a1185-cebc-4f89-88c2-63cfebe97759\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wfccb" Dec 01 09:16:14 crc kubenswrapper[4873]: I1201 09:16:14.095083 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89gv8\" (UniqueName: \"kubernetes.io/projected/cd2a1185-cebc-4f89-88c2-63cfebe97759-kube-api-access-89gv8\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wfccb\" (UID: \"cd2a1185-cebc-4f89-88c2-63cfebe97759\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wfccb" Dec 01 09:16:14 crc kubenswrapper[4873]: I1201 09:16:14.095347 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cd2a1185-cebc-4f89-88c2-63cfebe97759-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wfccb\" (UID: \"cd2a1185-cebc-4f89-88c2-63cfebe97759\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wfccb" Dec 01 09:16:14 crc kubenswrapper[4873]: I1201 09:16:14.095500 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cd2a1185-cebc-4f89-88c2-63cfebe97759-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wfccb\" (UID: \"cd2a1185-cebc-4f89-88c2-63cfebe97759\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wfccb" Dec 01 09:16:14 crc kubenswrapper[4873]: I1201 09:16:14.198434 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cd2a1185-cebc-4f89-88c2-63cfebe97759-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wfccb\" (UID: \"cd2a1185-cebc-4f89-88c2-63cfebe97759\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wfccb" Dec 01 09:16:14 crc kubenswrapper[4873]: I1201 09:16:14.198524 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd2a1185-cebc-4f89-88c2-63cfebe97759-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wfccb\" (UID: \"cd2a1185-cebc-4f89-88c2-63cfebe97759\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wfccb" Dec 01 09:16:14 crc kubenswrapper[4873]: I1201 09:16:14.198960 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89gv8\" (UniqueName: \"kubernetes.io/projected/cd2a1185-cebc-4f89-88c2-63cfebe97759-kube-api-access-89gv8\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wfccb\" (UID: \"cd2a1185-cebc-4f89-88c2-63cfebe97759\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wfccb" Dec 01 09:16:14 crc kubenswrapper[4873]: I1201 09:16:14.199243 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cd2a1185-cebc-4f89-88c2-63cfebe97759-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wfccb\" (UID: \"cd2a1185-cebc-4f89-88c2-63cfebe97759\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wfccb" Dec 01 09:16:14 crc kubenswrapper[4873]: I1201 09:16:14.199378 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cd2a1185-cebc-4f89-88c2-63cfebe97759-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wfccb\" (UID: \"cd2a1185-cebc-4f89-88c2-63cfebe97759\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wfccb" Dec 01 09:16:14 crc kubenswrapper[4873]: I1201 09:16:14.205685 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cd2a1185-cebc-4f89-88c2-63cfebe97759-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wfccb\" (UID: \"cd2a1185-cebc-4f89-88c2-63cfebe97759\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wfccb" Dec 01 09:16:14 crc kubenswrapper[4873]: I1201 09:16:14.206105 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd2a1185-cebc-4f89-88c2-63cfebe97759-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wfccb\" (UID: \"cd2a1185-cebc-4f89-88c2-63cfebe97759\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wfccb" Dec 01 09:16:14 crc kubenswrapper[4873]: I1201 09:16:14.209703 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cd2a1185-cebc-4f89-88c2-63cfebe97759-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wfccb\" (UID: \"cd2a1185-cebc-4f89-88c2-63cfebe97759\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wfccb" Dec 01 09:16:14 crc kubenswrapper[4873]: I1201 09:16:14.209951 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cd2a1185-cebc-4f89-88c2-63cfebe97759-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wfccb\" (UID: \"cd2a1185-cebc-4f89-88c2-63cfebe97759\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wfccb" Dec 01 09:16:14 crc kubenswrapper[4873]: I1201 09:16:14.228850 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89gv8\" (UniqueName: \"kubernetes.io/projected/cd2a1185-cebc-4f89-88c2-63cfebe97759-kube-api-access-89gv8\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wfccb\" (UID: \"cd2a1185-cebc-4f89-88c2-63cfebe97759\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wfccb" Dec 01 09:16:14 crc kubenswrapper[4873]: I1201 09:16:14.328584 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wfccb" Dec 01 09:16:15 crc kubenswrapper[4873]: I1201 09:16:15.025822 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wfccb"] Dec 01 09:16:15 crc kubenswrapper[4873]: W1201 09:16:15.030645 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd2a1185_cebc_4f89_88c2_63cfebe97759.slice/crio-7117128aaf609a98be60540f9d69bfdff4358feade008d8db8588c0e5391bfb1 WatchSource:0}: Error finding container 7117128aaf609a98be60540f9d69bfdff4358feade008d8db8588c0e5391bfb1: Status 404 returned error can't find the container with id 7117128aaf609a98be60540f9d69bfdff4358feade008d8db8588c0e5391bfb1 Dec 01 09:16:15 crc kubenswrapper[4873]: I1201 09:16:15.731971 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wfccb" event={"ID":"cd2a1185-cebc-4f89-88c2-63cfebe97759","Type":"ContainerStarted","Data":"7117128aaf609a98be60540f9d69bfdff4358feade008d8db8588c0e5391bfb1"} Dec 01 09:16:16 crc kubenswrapper[4873]: I1201 09:16:16.745192 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wfccb" event={"ID":"cd2a1185-cebc-4f89-88c2-63cfebe97759","Type":"ContainerStarted","Data":"24c1eb182b53b52b56b83093b868d7e8915318fe5cc8af06f81effeff6d5dca1"} Dec 01 09:16:16 crc kubenswrapper[4873]: I1201 09:16:16.782371 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wfccb" podStartSLOduration=3.284142976 podStartE2EDuration="3.782348848s" podCreationTimestamp="2025-12-01 09:16:13 +0000 UTC" firstStartedPulling="2025-12-01 09:16:15.033613342 +0000 UTC m=+2150.935721881" lastFinishedPulling="2025-12-01 09:16:15.531819214 +0000 UTC m=+2151.433927753" observedRunningTime="2025-12-01 09:16:16.776780991 +0000 UTC m=+2152.678889550" watchObservedRunningTime="2025-12-01 09:16:16.782348848 +0000 UTC m=+2152.684457387" Dec 01 09:16:31 crc kubenswrapper[4873]: I1201 09:16:31.058904 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:16:31 crc kubenswrapper[4873]: I1201 09:16:31.059632 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:16:31 crc kubenswrapper[4873]: I1201 09:16:31.737630 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kd7nt"] Dec 01 09:16:31 crc kubenswrapper[4873]: I1201 09:16:31.740103 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kd7nt" Dec 01 09:16:31 crc kubenswrapper[4873]: I1201 09:16:31.756411 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kd7nt"] Dec 01 09:16:31 crc kubenswrapper[4873]: I1201 09:16:31.805826 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d99932d2-5e5c-44ed-afa1-6f7616f5a03b-catalog-content\") pod \"redhat-marketplace-kd7nt\" (UID: \"d99932d2-5e5c-44ed-afa1-6f7616f5a03b\") " pod="openshift-marketplace/redhat-marketplace-kd7nt" Dec 01 09:16:31 crc kubenswrapper[4873]: I1201 09:16:31.806003 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cl74\" (UniqueName: \"kubernetes.io/projected/d99932d2-5e5c-44ed-afa1-6f7616f5a03b-kube-api-access-8cl74\") pod \"redhat-marketplace-kd7nt\" (UID: \"d99932d2-5e5c-44ed-afa1-6f7616f5a03b\") " pod="openshift-marketplace/redhat-marketplace-kd7nt" Dec 01 09:16:31 crc kubenswrapper[4873]: I1201 09:16:31.806042 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d99932d2-5e5c-44ed-afa1-6f7616f5a03b-utilities\") pod \"redhat-marketplace-kd7nt\" (UID: \"d99932d2-5e5c-44ed-afa1-6f7616f5a03b\") " pod="openshift-marketplace/redhat-marketplace-kd7nt" Dec 01 09:16:31 crc kubenswrapper[4873]: I1201 09:16:31.937924 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cl74\" (UniqueName: \"kubernetes.io/projected/d99932d2-5e5c-44ed-afa1-6f7616f5a03b-kube-api-access-8cl74\") pod \"redhat-marketplace-kd7nt\" (UID: \"d99932d2-5e5c-44ed-afa1-6f7616f5a03b\") " pod="openshift-marketplace/redhat-marketplace-kd7nt" Dec 01 09:16:31 crc kubenswrapper[4873]: I1201 09:16:31.938041 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d99932d2-5e5c-44ed-afa1-6f7616f5a03b-utilities\") pod \"redhat-marketplace-kd7nt\" (UID: \"d99932d2-5e5c-44ed-afa1-6f7616f5a03b\") " pod="openshift-marketplace/redhat-marketplace-kd7nt" Dec 01 09:16:31 crc kubenswrapper[4873]: I1201 09:16:31.938385 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d99932d2-5e5c-44ed-afa1-6f7616f5a03b-catalog-content\") pod \"redhat-marketplace-kd7nt\" (UID: \"d99932d2-5e5c-44ed-afa1-6f7616f5a03b\") " pod="openshift-marketplace/redhat-marketplace-kd7nt" Dec 01 09:16:31 crc kubenswrapper[4873]: I1201 09:16:31.938947 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d99932d2-5e5c-44ed-afa1-6f7616f5a03b-utilities\") pod \"redhat-marketplace-kd7nt\" (UID: \"d99932d2-5e5c-44ed-afa1-6f7616f5a03b\") " pod="openshift-marketplace/redhat-marketplace-kd7nt" Dec 01 09:16:31 crc kubenswrapper[4873]: I1201 09:16:31.939116 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d99932d2-5e5c-44ed-afa1-6f7616f5a03b-catalog-content\") pod \"redhat-marketplace-kd7nt\" (UID: \"d99932d2-5e5c-44ed-afa1-6f7616f5a03b\") " pod="openshift-marketplace/redhat-marketplace-kd7nt" Dec 01 09:16:31 crc kubenswrapper[4873]: I1201 09:16:31.964529 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cl74\" (UniqueName: \"kubernetes.io/projected/d99932d2-5e5c-44ed-afa1-6f7616f5a03b-kube-api-access-8cl74\") pod \"redhat-marketplace-kd7nt\" (UID: \"d99932d2-5e5c-44ed-afa1-6f7616f5a03b\") " pod="openshift-marketplace/redhat-marketplace-kd7nt" Dec 01 09:16:32 crc kubenswrapper[4873]: I1201 09:16:32.063091 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kd7nt" Dec 01 09:16:32 crc kubenswrapper[4873]: I1201 09:16:32.554418 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kd7nt"] Dec 01 09:16:32 crc kubenswrapper[4873]: I1201 09:16:32.953132 4873 generic.go:334] "Generic (PLEG): container finished" podID="d99932d2-5e5c-44ed-afa1-6f7616f5a03b" containerID="a0006813c27bed651a0839054cdb4b949d00f7a165bce0213bf1cf5f37e9cee7" exitCode=0 Dec 01 09:16:32 crc kubenswrapper[4873]: I1201 09:16:32.953202 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kd7nt" event={"ID":"d99932d2-5e5c-44ed-afa1-6f7616f5a03b","Type":"ContainerDied","Data":"a0006813c27bed651a0839054cdb4b949d00f7a165bce0213bf1cf5f37e9cee7"} Dec 01 09:16:32 crc kubenswrapper[4873]: I1201 09:16:32.953529 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kd7nt" event={"ID":"d99932d2-5e5c-44ed-afa1-6f7616f5a03b","Type":"ContainerStarted","Data":"8f626312ad260e82cb2472242d90a91178a873cbb2fdd87f2607f1c1934b9009"} Dec 01 09:16:34 crc kubenswrapper[4873]: I1201 09:16:34.976060 4873 generic.go:334] "Generic (PLEG): container finished" podID="d99932d2-5e5c-44ed-afa1-6f7616f5a03b" containerID="d66cf77dc5d9cc08cd7e4b7e12c768e2edb046d7cd9122e53fbce5b1ce5f9548" exitCode=0 Dec 01 09:16:34 crc kubenswrapper[4873]: I1201 09:16:34.976280 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kd7nt" event={"ID":"d99932d2-5e5c-44ed-afa1-6f7616f5a03b","Type":"ContainerDied","Data":"d66cf77dc5d9cc08cd7e4b7e12c768e2edb046d7cd9122e53fbce5b1ce5f9548"} Dec 01 09:16:35 crc kubenswrapper[4873]: I1201 09:16:35.990318 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kd7nt" event={"ID":"d99932d2-5e5c-44ed-afa1-6f7616f5a03b","Type":"ContainerStarted","Data":"5556e8240fe6651d9258b362f7f14308fe6735bb2c839d2d7dac54b7170ac04c"} Dec 01 09:16:36 crc kubenswrapper[4873]: I1201 09:16:36.012925 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kd7nt" podStartSLOduration=2.52071192 podStartE2EDuration="5.012903601s" podCreationTimestamp="2025-12-01 09:16:31 +0000 UTC" firstStartedPulling="2025-12-01 09:16:32.95531192 +0000 UTC m=+2168.857420459" lastFinishedPulling="2025-12-01 09:16:35.447503591 +0000 UTC m=+2171.349612140" observedRunningTime="2025-12-01 09:16:36.008329368 +0000 UTC m=+2171.910437907" watchObservedRunningTime="2025-12-01 09:16:36.012903601 +0000 UTC m=+2171.915012140" Dec 01 09:16:40 crc kubenswrapper[4873]: I1201 09:16:40.019735 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rd2ml"] Dec 01 09:16:40 crc kubenswrapper[4873]: I1201 09:16:40.024104 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rd2ml" Dec 01 09:16:40 crc kubenswrapper[4873]: I1201 09:16:40.041847 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rd2ml"] Dec 01 09:16:40 crc kubenswrapper[4873]: I1201 09:16:40.117678 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20057a0c-212d-4154-9556-dfbbf835c625-utilities\") pod \"redhat-operators-rd2ml\" (UID: \"20057a0c-212d-4154-9556-dfbbf835c625\") " pod="openshift-marketplace/redhat-operators-rd2ml" Dec 01 09:16:40 crc kubenswrapper[4873]: I1201 09:16:40.118045 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s89wp\" (UniqueName: \"kubernetes.io/projected/20057a0c-212d-4154-9556-dfbbf835c625-kube-api-access-s89wp\") pod \"redhat-operators-rd2ml\" (UID: \"20057a0c-212d-4154-9556-dfbbf835c625\") " pod="openshift-marketplace/redhat-operators-rd2ml" Dec 01 09:16:40 crc kubenswrapper[4873]: I1201 09:16:40.118322 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20057a0c-212d-4154-9556-dfbbf835c625-catalog-content\") pod \"redhat-operators-rd2ml\" (UID: \"20057a0c-212d-4154-9556-dfbbf835c625\") " pod="openshift-marketplace/redhat-operators-rd2ml" Dec 01 09:16:40 crc kubenswrapper[4873]: I1201 09:16:40.220740 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s89wp\" (UniqueName: \"kubernetes.io/projected/20057a0c-212d-4154-9556-dfbbf835c625-kube-api-access-s89wp\") pod \"redhat-operators-rd2ml\" (UID: \"20057a0c-212d-4154-9556-dfbbf835c625\") " pod="openshift-marketplace/redhat-operators-rd2ml" Dec 01 09:16:40 crc kubenswrapper[4873]: I1201 09:16:40.220817 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20057a0c-212d-4154-9556-dfbbf835c625-catalog-content\") pod \"redhat-operators-rd2ml\" (UID: \"20057a0c-212d-4154-9556-dfbbf835c625\") " pod="openshift-marketplace/redhat-operators-rd2ml" Dec 01 09:16:40 crc kubenswrapper[4873]: I1201 09:16:40.220914 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20057a0c-212d-4154-9556-dfbbf835c625-utilities\") pod \"redhat-operators-rd2ml\" (UID: \"20057a0c-212d-4154-9556-dfbbf835c625\") " pod="openshift-marketplace/redhat-operators-rd2ml" Dec 01 09:16:40 crc kubenswrapper[4873]: I1201 09:16:40.221435 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20057a0c-212d-4154-9556-dfbbf835c625-utilities\") pod \"redhat-operators-rd2ml\" (UID: \"20057a0c-212d-4154-9556-dfbbf835c625\") " pod="openshift-marketplace/redhat-operators-rd2ml" Dec 01 09:16:40 crc kubenswrapper[4873]: I1201 09:16:40.221559 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20057a0c-212d-4154-9556-dfbbf835c625-catalog-content\") pod \"redhat-operators-rd2ml\" (UID: \"20057a0c-212d-4154-9556-dfbbf835c625\") " pod="openshift-marketplace/redhat-operators-rd2ml" Dec 01 09:16:40 crc kubenswrapper[4873]: I1201 09:16:40.245053 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s89wp\" (UniqueName: \"kubernetes.io/projected/20057a0c-212d-4154-9556-dfbbf835c625-kube-api-access-s89wp\") pod \"redhat-operators-rd2ml\" (UID: \"20057a0c-212d-4154-9556-dfbbf835c625\") " pod="openshift-marketplace/redhat-operators-rd2ml" Dec 01 09:16:40 crc kubenswrapper[4873]: I1201 09:16:40.347589 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rd2ml" Dec 01 09:16:40 crc kubenswrapper[4873]: I1201 09:16:40.932923 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rd2ml"] Dec 01 09:16:41 crc kubenswrapper[4873]: I1201 09:16:41.051191 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rd2ml" event={"ID":"20057a0c-212d-4154-9556-dfbbf835c625","Type":"ContainerStarted","Data":"367ff0671e55c2f11b8ccf5b510d7367ebd8a450dfdb7e3cabe60982c9288584"} Dec 01 09:16:42 crc kubenswrapper[4873]: I1201 09:16:42.064241 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kd7nt" Dec 01 09:16:42 crc kubenswrapper[4873]: I1201 09:16:42.064855 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kd7nt" Dec 01 09:16:42 crc kubenswrapper[4873]: I1201 09:16:42.065610 4873 generic.go:334] "Generic (PLEG): container finished" podID="20057a0c-212d-4154-9556-dfbbf835c625" containerID="d1ff72b15bad8d1f83737a9ddcc70e1eac35db6264e3e04e97b55d679f040f03" exitCode=0 Dec 01 09:16:42 crc kubenswrapper[4873]: I1201 09:16:42.065660 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rd2ml" event={"ID":"20057a0c-212d-4154-9556-dfbbf835c625","Type":"ContainerDied","Data":"d1ff72b15bad8d1f83737a9ddcc70e1eac35db6264e3e04e97b55d679f040f03"} Dec 01 09:16:42 crc kubenswrapper[4873]: I1201 09:16:42.127941 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kd7nt" Dec 01 09:16:43 crc kubenswrapper[4873]: I1201 09:16:43.078222 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rd2ml" event={"ID":"20057a0c-212d-4154-9556-dfbbf835c625","Type":"ContainerStarted","Data":"630bfd3e8daae0cae9e29c55c01b1e082a20af3583a23c19b2051756dd86ab7e"} Dec 01 09:16:43 crc kubenswrapper[4873]: I1201 09:16:43.136970 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kd7nt" Dec 01 09:16:44 crc kubenswrapper[4873]: I1201 09:16:44.394165 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kd7nt"] Dec 01 09:16:45 crc kubenswrapper[4873]: I1201 09:16:45.102574 4873 generic.go:334] "Generic (PLEG): container finished" podID="20057a0c-212d-4154-9556-dfbbf835c625" containerID="630bfd3e8daae0cae9e29c55c01b1e082a20af3583a23c19b2051756dd86ab7e" exitCode=0 Dec 01 09:16:45 crc kubenswrapper[4873]: I1201 09:16:45.102676 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rd2ml" event={"ID":"20057a0c-212d-4154-9556-dfbbf835c625","Type":"ContainerDied","Data":"630bfd3e8daae0cae9e29c55c01b1e082a20af3583a23c19b2051756dd86ab7e"} Dec 01 09:16:45 crc kubenswrapper[4873]: I1201 09:16:45.103321 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kd7nt" podUID="d99932d2-5e5c-44ed-afa1-6f7616f5a03b" containerName="registry-server" containerID="cri-o://5556e8240fe6651d9258b362f7f14308fe6735bb2c839d2d7dac54b7170ac04c" gracePeriod=2 Dec 01 09:16:45 crc kubenswrapper[4873]: I1201 09:16:45.620583 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kd7nt" Dec 01 09:16:45 crc kubenswrapper[4873]: I1201 09:16:45.645764 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d99932d2-5e5c-44ed-afa1-6f7616f5a03b-utilities\") pod \"d99932d2-5e5c-44ed-afa1-6f7616f5a03b\" (UID: \"d99932d2-5e5c-44ed-afa1-6f7616f5a03b\") " Dec 01 09:16:45 crc kubenswrapper[4873]: I1201 09:16:45.645878 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8cl74\" (UniqueName: \"kubernetes.io/projected/d99932d2-5e5c-44ed-afa1-6f7616f5a03b-kube-api-access-8cl74\") pod \"d99932d2-5e5c-44ed-afa1-6f7616f5a03b\" (UID: \"d99932d2-5e5c-44ed-afa1-6f7616f5a03b\") " Dec 01 09:16:45 crc kubenswrapper[4873]: I1201 09:16:45.649889 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d99932d2-5e5c-44ed-afa1-6f7616f5a03b-utilities" (OuterVolumeSpecName: "utilities") pod "d99932d2-5e5c-44ed-afa1-6f7616f5a03b" (UID: "d99932d2-5e5c-44ed-afa1-6f7616f5a03b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:16:45 crc kubenswrapper[4873]: I1201 09:16:45.655229 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d99932d2-5e5c-44ed-afa1-6f7616f5a03b-kube-api-access-8cl74" (OuterVolumeSpecName: "kube-api-access-8cl74") pod "d99932d2-5e5c-44ed-afa1-6f7616f5a03b" (UID: "d99932d2-5e5c-44ed-afa1-6f7616f5a03b"). InnerVolumeSpecName "kube-api-access-8cl74". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:16:45 crc kubenswrapper[4873]: I1201 09:16:45.748630 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d99932d2-5e5c-44ed-afa1-6f7616f5a03b-catalog-content\") pod \"d99932d2-5e5c-44ed-afa1-6f7616f5a03b\" (UID: \"d99932d2-5e5c-44ed-afa1-6f7616f5a03b\") " Dec 01 09:16:45 crc kubenswrapper[4873]: I1201 09:16:45.749695 4873 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d99932d2-5e5c-44ed-afa1-6f7616f5a03b-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 09:16:45 crc kubenswrapper[4873]: I1201 09:16:45.749723 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8cl74\" (UniqueName: \"kubernetes.io/projected/d99932d2-5e5c-44ed-afa1-6f7616f5a03b-kube-api-access-8cl74\") on node \"crc\" DevicePath \"\"" Dec 01 09:16:45 crc kubenswrapper[4873]: I1201 09:16:45.777847 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d99932d2-5e5c-44ed-afa1-6f7616f5a03b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d99932d2-5e5c-44ed-afa1-6f7616f5a03b" (UID: "d99932d2-5e5c-44ed-afa1-6f7616f5a03b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:16:45 crc kubenswrapper[4873]: I1201 09:16:45.852819 4873 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d99932d2-5e5c-44ed-afa1-6f7616f5a03b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 09:16:46 crc kubenswrapper[4873]: I1201 09:16:46.118163 4873 generic.go:334] "Generic (PLEG): container finished" podID="d99932d2-5e5c-44ed-afa1-6f7616f5a03b" containerID="5556e8240fe6651d9258b362f7f14308fe6735bb2c839d2d7dac54b7170ac04c" exitCode=0 Dec 01 09:16:46 crc kubenswrapper[4873]: I1201 09:16:46.118229 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kd7nt" event={"ID":"d99932d2-5e5c-44ed-afa1-6f7616f5a03b","Type":"ContainerDied","Data":"5556e8240fe6651d9258b362f7f14308fe6735bb2c839d2d7dac54b7170ac04c"} Dec 01 09:16:46 crc kubenswrapper[4873]: I1201 09:16:46.118245 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kd7nt" Dec 01 09:16:46 crc kubenswrapper[4873]: I1201 09:16:46.118270 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kd7nt" event={"ID":"d99932d2-5e5c-44ed-afa1-6f7616f5a03b","Type":"ContainerDied","Data":"8f626312ad260e82cb2472242d90a91178a873cbb2fdd87f2607f1c1934b9009"} Dec 01 09:16:46 crc kubenswrapper[4873]: I1201 09:16:46.118322 4873 scope.go:117] "RemoveContainer" containerID="5556e8240fe6651d9258b362f7f14308fe6735bb2c839d2d7dac54b7170ac04c" Dec 01 09:16:46 crc kubenswrapper[4873]: I1201 09:16:46.150913 4873 scope.go:117] "RemoveContainer" containerID="d66cf77dc5d9cc08cd7e4b7e12c768e2edb046d7cd9122e53fbce5b1ce5f9548" Dec 01 09:16:46 crc kubenswrapper[4873]: I1201 09:16:46.163582 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kd7nt"] Dec 01 09:16:46 crc kubenswrapper[4873]: I1201 09:16:46.173876 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kd7nt"] Dec 01 09:16:46 crc kubenswrapper[4873]: I1201 09:16:46.177965 4873 scope.go:117] "RemoveContainer" containerID="a0006813c27bed651a0839054cdb4b949d00f7a165bce0213bf1cf5f37e9cee7" Dec 01 09:16:46 crc kubenswrapper[4873]: I1201 09:16:46.222679 4873 scope.go:117] "RemoveContainer" containerID="5556e8240fe6651d9258b362f7f14308fe6735bb2c839d2d7dac54b7170ac04c" Dec 01 09:16:46 crc kubenswrapper[4873]: E1201 09:16:46.223532 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5556e8240fe6651d9258b362f7f14308fe6735bb2c839d2d7dac54b7170ac04c\": container with ID starting with 5556e8240fe6651d9258b362f7f14308fe6735bb2c839d2d7dac54b7170ac04c not found: ID does not exist" containerID="5556e8240fe6651d9258b362f7f14308fe6735bb2c839d2d7dac54b7170ac04c" Dec 01 09:16:46 crc kubenswrapper[4873]: I1201 09:16:46.223574 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5556e8240fe6651d9258b362f7f14308fe6735bb2c839d2d7dac54b7170ac04c"} err="failed to get container status \"5556e8240fe6651d9258b362f7f14308fe6735bb2c839d2d7dac54b7170ac04c\": rpc error: code = NotFound desc = could not find container \"5556e8240fe6651d9258b362f7f14308fe6735bb2c839d2d7dac54b7170ac04c\": container with ID starting with 5556e8240fe6651d9258b362f7f14308fe6735bb2c839d2d7dac54b7170ac04c not found: ID does not exist" Dec 01 09:16:46 crc kubenswrapper[4873]: I1201 09:16:46.223605 4873 scope.go:117] "RemoveContainer" containerID="d66cf77dc5d9cc08cd7e4b7e12c768e2edb046d7cd9122e53fbce5b1ce5f9548" Dec 01 09:16:46 crc kubenswrapper[4873]: E1201 09:16:46.224114 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d66cf77dc5d9cc08cd7e4b7e12c768e2edb046d7cd9122e53fbce5b1ce5f9548\": container with ID starting with d66cf77dc5d9cc08cd7e4b7e12c768e2edb046d7cd9122e53fbce5b1ce5f9548 not found: ID does not exist" containerID="d66cf77dc5d9cc08cd7e4b7e12c768e2edb046d7cd9122e53fbce5b1ce5f9548" Dec 01 09:16:46 crc kubenswrapper[4873]: I1201 09:16:46.224170 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d66cf77dc5d9cc08cd7e4b7e12c768e2edb046d7cd9122e53fbce5b1ce5f9548"} err="failed to get container status \"d66cf77dc5d9cc08cd7e4b7e12c768e2edb046d7cd9122e53fbce5b1ce5f9548\": rpc error: code = NotFound desc = could not find container \"d66cf77dc5d9cc08cd7e4b7e12c768e2edb046d7cd9122e53fbce5b1ce5f9548\": container with ID starting with d66cf77dc5d9cc08cd7e4b7e12c768e2edb046d7cd9122e53fbce5b1ce5f9548 not found: ID does not exist" Dec 01 09:16:46 crc kubenswrapper[4873]: I1201 09:16:46.224208 4873 scope.go:117] "RemoveContainer" containerID="a0006813c27bed651a0839054cdb4b949d00f7a165bce0213bf1cf5f37e9cee7" Dec 01 09:16:46 crc kubenswrapper[4873]: E1201 09:16:46.224566 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0006813c27bed651a0839054cdb4b949d00f7a165bce0213bf1cf5f37e9cee7\": container with ID starting with a0006813c27bed651a0839054cdb4b949d00f7a165bce0213bf1cf5f37e9cee7 not found: ID does not exist" containerID="a0006813c27bed651a0839054cdb4b949d00f7a165bce0213bf1cf5f37e9cee7" Dec 01 09:16:46 crc kubenswrapper[4873]: I1201 09:16:46.224608 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0006813c27bed651a0839054cdb4b949d00f7a165bce0213bf1cf5f37e9cee7"} err="failed to get container status \"a0006813c27bed651a0839054cdb4b949d00f7a165bce0213bf1cf5f37e9cee7\": rpc error: code = NotFound desc = could not find container \"a0006813c27bed651a0839054cdb4b949d00f7a165bce0213bf1cf5f37e9cee7\": container with ID starting with a0006813c27bed651a0839054cdb4b949d00f7a165bce0213bf1cf5f37e9cee7 not found: ID does not exist" Dec 01 09:16:46 crc kubenswrapper[4873]: I1201 09:16:46.443934 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d99932d2-5e5c-44ed-afa1-6f7616f5a03b" path="/var/lib/kubelet/pods/d99932d2-5e5c-44ed-afa1-6f7616f5a03b/volumes" Dec 01 09:16:47 crc kubenswrapper[4873]: I1201 09:16:47.133636 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rd2ml" event={"ID":"20057a0c-212d-4154-9556-dfbbf835c625","Type":"ContainerStarted","Data":"00ae366723b3e59bb1e0e0192a3537a052682d90600b2c078491f160981c3550"} Dec 01 09:16:47 crc kubenswrapper[4873]: I1201 09:16:47.161077 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rd2ml" podStartSLOduration=3.989450636 podStartE2EDuration="8.161052334s" podCreationTimestamp="2025-12-01 09:16:39 +0000 UTC" firstStartedPulling="2025-12-01 09:16:42.068671327 +0000 UTC m=+2177.970779856" lastFinishedPulling="2025-12-01 09:16:46.240273015 +0000 UTC m=+2182.142381554" observedRunningTime="2025-12-01 09:16:47.157431455 +0000 UTC m=+2183.059539994" watchObservedRunningTime="2025-12-01 09:16:47.161052334 +0000 UTC m=+2183.063160873" Dec 01 09:16:50 crc kubenswrapper[4873]: I1201 09:16:50.347973 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rd2ml" Dec 01 09:16:50 crc kubenswrapper[4873]: I1201 09:16:50.348687 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rd2ml" Dec 01 09:16:51 crc kubenswrapper[4873]: I1201 09:16:51.404947 4873 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rd2ml" podUID="20057a0c-212d-4154-9556-dfbbf835c625" containerName="registry-server" probeResult="failure" output=< Dec 01 09:16:51 crc kubenswrapper[4873]: timeout: failed to connect service ":50051" within 1s Dec 01 09:16:51 crc kubenswrapper[4873]: > Dec 01 09:16:58 crc kubenswrapper[4873]: I1201 09:16:58.368975 4873 scope.go:117] "RemoveContainer" containerID="0950a0e43d01de59e4c5d634f772f6405eefb78ed37e32db1c3763b4722e6a80" Dec 01 09:17:00 crc kubenswrapper[4873]: I1201 09:17:00.395885 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rd2ml" Dec 01 09:17:00 crc kubenswrapper[4873]: I1201 09:17:00.446101 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rd2ml" Dec 01 09:17:00 crc kubenswrapper[4873]: I1201 09:17:00.635592 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rd2ml"] Dec 01 09:17:01 crc kubenswrapper[4873]: I1201 09:17:01.059390 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:17:01 crc kubenswrapper[4873]: I1201 09:17:01.059480 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:17:01 crc kubenswrapper[4873]: I1201 09:17:01.059541 4873 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" Dec 01 09:17:01 crc kubenswrapper[4873]: I1201 09:17:01.060674 4873 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1a1db8f8327e8cf6dbb94b57ea898b250339bb198ae1ba595cb1d79805fa0db1"} pod="openshift-machine-config-operator/machine-config-daemon-scwpp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 09:17:01 crc kubenswrapper[4873]: I1201 09:17:01.060756 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" containerID="cri-o://1a1db8f8327e8cf6dbb94b57ea898b250339bb198ae1ba595cb1d79805fa0db1" gracePeriod=600 Dec 01 09:17:01 crc kubenswrapper[4873]: E1201 09:17:01.186993 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:17:01 crc kubenswrapper[4873]: I1201 09:17:01.291924 4873 generic.go:334] "Generic (PLEG): container finished" podID="fef7b114-0e07-402d-a37b-315c36011f4b" containerID="1a1db8f8327e8cf6dbb94b57ea898b250339bb198ae1ba595cb1d79805fa0db1" exitCode=0 Dec 01 09:17:01 crc kubenswrapper[4873]: I1201 09:17:01.292646 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" event={"ID":"fef7b114-0e07-402d-a37b-315c36011f4b","Type":"ContainerDied","Data":"1a1db8f8327e8cf6dbb94b57ea898b250339bb198ae1ba595cb1d79805fa0db1"} Dec 01 09:17:01 crc kubenswrapper[4873]: I1201 09:17:01.292713 4873 scope.go:117] "RemoveContainer" containerID="281f1071cdd39ef7daaacd7ebb1e90ad778c28e93bebe102e714733c9cf3c602" Dec 01 09:17:01 crc kubenswrapper[4873]: I1201 09:17:01.294097 4873 scope.go:117] "RemoveContainer" containerID="1a1db8f8327e8cf6dbb94b57ea898b250339bb198ae1ba595cb1d79805fa0db1" Dec 01 09:17:01 crc kubenswrapper[4873]: E1201 09:17:01.294464 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:17:02 crc kubenswrapper[4873]: I1201 09:17:02.305526 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rd2ml" podUID="20057a0c-212d-4154-9556-dfbbf835c625" containerName="registry-server" containerID="cri-o://00ae366723b3e59bb1e0e0192a3537a052682d90600b2c078491f160981c3550" gracePeriod=2 Dec 01 09:17:02 crc kubenswrapper[4873]: I1201 09:17:02.843574 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rd2ml" Dec 01 09:17:02 crc kubenswrapper[4873]: I1201 09:17:02.913584 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s89wp\" (UniqueName: \"kubernetes.io/projected/20057a0c-212d-4154-9556-dfbbf835c625-kube-api-access-s89wp\") pod \"20057a0c-212d-4154-9556-dfbbf835c625\" (UID: \"20057a0c-212d-4154-9556-dfbbf835c625\") " Dec 01 09:17:02 crc kubenswrapper[4873]: I1201 09:17:02.914180 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20057a0c-212d-4154-9556-dfbbf835c625-catalog-content\") pod \"20057a0c-212d-4154-9556-dfbbf835c625\" (UID: \"20057a0c-212d-4154-9556-dfbbf835c625\") " Dec 01 09:17:02 crc kubenswrapper[4873]: I1201 09:17:02.914318 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20057a0c-212d-4154-9556-dfbbf835c625-utilities\") pod \"20057a0c-212d-4154-9556-dfbbf835c625\" (UID: \"20057a0c-212d-4154-9556-dfbbf835c625\") " Dec 01 09:17:02 crc kubenswrapper[4873]: I1201 09:17:02.915057 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20057a0c-212d-4154-9556-dfbbf835c625-utilities" (OuterVolumeSpecName: "utilities") pod "20057a0c-212d-4154-9556-dfbbf835c625" (UID: "20057a0c-212d-4154-9556-dfbbf835c625"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:17:02 crc kubenswrapper[4873]: I1201 09:17:02.921109 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20057a0c-212d-4154-9556-dfbbf835c625-kube-api-access-s89wp" (OuterVolumeSpecName: "kube-api-access-s89wp") pod "20057a0c-212d-4154-9556-dfbbf835c625" (UID: "20057a0c-212d-4154-9556-dfbbf835c625"). InnerVolumeSpecName "kube-api-access-s89wp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:17:03 crc kubenswrapper[4873]: I1201 09:17:03.019722 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s89wp\" (UniqueName: \"kubernetes.io/projected/20057a0c-212d-4154-9556-dfbbf835c625-kube-api-access-s89wp\") on node \"crc\" DevicePath \"\"" Dec 01 09:17:03 crc kubenswrapper[4873]: I1201 09:17:03.019781 4873 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20057a0c-212d-4154-9556-dfbbf835c625-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 09:17:03 crc kubenswrapper[4873]: I1201 09:17:03.038482 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20057a0c-212d-4154-9556-dfbbf835c625-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "20057a0c-212d-4154-9556-dfbbf835c625" (UID: "20057a0c-212d-4154-9556-dfbbf835c625"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:17:03 crc kubenswrapper[4873]: I1201 09:17:03.121854 4873 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20057a0c-212d-4154-9556-dfbbf835c625-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 09:17:03 crc kubenswrapper[4873]: I1201 09:17:03.320560 4873 generic.go:334] "Generic (PLEG): container finished" podID="20057a0c-212d-4154-9556-dfbbf835c625" containerID="00ae366723b3e59bb1e0e0192a3537a052682d90600b2c078491f160981c3550" exitCode=0 Dec 01 09:17:03 crc kubenswrapper[4873]: I1201 09:17:03.320662 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rd2ml" Dec 01 09:17:03 crc kubenswrapper[4873]: I1201 09:17:03.320695 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rd2ml" event={"ID":"20057a0c-212d-4154-9556-dfbbf835c625","Type":"ContainerDied","Data":"00ae366723b3e59bb1e0e0192a3537a052682d90600b2c078491f160981c3550"} Dec 01 09:17:03 crc kubenswrapper[4873]: I1201 09:17:03.321205 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rd2ml" event={"ID":"20057a0c-212d-4154-9556-dfbbf835c625","Type":"ContainerDied","Data":"367ff0671e55c2f11b8ccf5b510d7367ebd8a450dfdb7e3cabe60982c9288584"} Dec 01 09:17:03 crc kubenswrapper[4873]: I1201 09:17:03.321228 4873 scope.go:117] "RemoveContainer" containerID="00ae366723b3e59bb1e0e0192a3537a052682d90600b2c078491f160981c3550" Dec 01 09:17:03 crc kubenswrapper[4873]: I1201 09:17:03.349478 4873 scope.go:117] "RemoveContainer" containerID="630bfd3e8daae0cae9e29c55c01b1e082a20af3583a23c19b2051756dd86ab7e" Dec 01 09:17:03 crc kubenswrapper[4873]: I1201 09:17:03.372685 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rd2ml"] Dec 01 09:17:03 crc kubenswrapper[4873]: I1201 09:17:03.383651 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rd2ml"] Dec 01 09:17:03 crc kubenswrapper[4873]: I1201 09:17:03.396903 4873 scope.go:117] "RemoveContainer" containerID="d1ff72b15bad8d1f83737a9ddcc70e1eac35db6264e3e04e97b55d679f040f03" Dec 01 09:17:03 crc kubenswrapper[4873]: I1201 09:17:03.432214 4873 scope.go:117] "RemoveContainer" containerID="00ae366723b3e59bb1e0e0192a3537a052682d90600b2c078491f160981c3550" Dec 01 09:17:03 crc kubenswrapper[4873]: E1201 09:17:03.432972 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00ae366723b3e59bb1e0e0192a3537a052682d90600b2c078491f160981c3550\": container with ID starting with 00ae366723b3e59bb1e0e0192a3537a052682d90600b2c078491f160981c3550 not found: ID does not exist" containerID="00ae366723b3e59bb1e0e0192a3537a052682d90600b2c078491f160981c3550" Dec 01 09:17:03 crc kubenswrapper[4873]: I1201 09:17:03.433031 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00ae366723b3e59bb1e0e0192a3537a052682d90600b2c078491f160981c3550"} err="failed to get container status \"00ae366723b3e59bb1e0e0192a3537a052682d90600b2c078491f160981c3550\": rpc error: code = NotFound desc = could not find container \"00ae366723b3e59bb1e0e0192a3537a052682d90600b2c078491f160981c3550\": container with ID starting with 00ae366723b3e59bb1e0e0192a3537a052682d90600b2c078491f160981c3550 not found: ID does not exist" Dec 01 09:17:03 crc kubenswrapper[4873]: I1201 09:17:03.433075 4873 scope.go:117] "RemoveContainer" containerID="630bfd3e8daae0cae9e29c55c01b1e082a20af3583a23c19b2051756dd86ab7e" Dec 01 09:17:03 crc kubenswrapper[4873]: E1201 09:17:03.433936 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"630bfd3e8daae0cae9e29c55c01b1e082a20af3583a23c19b2051756dd86ab7e\": container with ID starting with 630bfd3e8daae0cae9e29c55c01b1e082a20af3583a23c19b2051756dd86ab7e not found: ID does not exist" containerID="630bfd3e8daae0cae9e29c55c01b1e082a20af3583a23c19b2051756dd86ab7e" Dec 01 09:17:03 crc kubenswrapper[4873]: I1201 09:17:03.434119 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"630bfd3e8daae0cae9e29c55c01b1e082a20af3583a23c19b2051756dd86ab7e"} err="failed to get container status \"630bfd3e8daae0cae9e29c55c01b1e082a20af3583a23c19b2051756dd86ab7e\": rpc error: code = NotFound desc = could not find container \"630bfd3e8daae0cae9e29c55c01b1e082a20af3583a23c19b2051756dd86ab7e\": container with ID starting with 630bfd3e8daae0cae9e29c55c01b1e082a20af3583a23c19b2051756dd86ab7e not found: ID does not exist" Dec 01 09:17:03 crc kubenswrapper[4873]: I1201 09:17:03.434222 4873 scope.go:117] "RemoveContainer" containerID="d1ff72b15bad8d1f83737a9ddcc70e1eac35db6264e3e04e97b55d679f040f03" Dec 01 09:17:03 crc kubenswrapper[4873]: E1201 09:17:03.435059 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1ff72b15bad8d1f83737a9ddcc70e1eac35db6264e3e04e97b55d679f040f03\": container with ID starting with d1ff72b15bad8d1f83737a9ddcc70e1eac35db6264e3e04e97b55d679f040f03 not found: ID does not exist" containerID="d1ff72b15bad8d1f83737a9ddcc70e1eac35db6264e3e04e97b55d679f040f03" Dec 01 09:17:03 crc kubenswrapper[4873]: I1201 09:17:03.435115 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1ff72b15bad8d1f83737a9ddcc70e1eac35db6264e3e04e97b55d679f040f03"} err="failed to get container status \"d1ff72b15bad8d1f83737a9ddcc70e1eac35db6264e3e04e97b55d679f040f03\": rpc error: code = NotFound desc = could not find container \"d1ff72b15bad8d1f83737a9ddcc70e1eac35db6264e3e04e97b55d679f040f03\": container with ID starting with d1ff72b15bad8d1f83737a9ddcc70e1eac35db6264e3e04e97b55d679f040f03 not found: ID does not exist" Dec 01 09:17:04 crc kubenswrapper[4873]: I1201 09:17:04.450197 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20057a0c-212d-4154-9556-dfbbf835c625" path="/var/lib/kubelet/pods/20057a0c-212d-4154-9556-dfbbf835c625/volumes" Dec 01 09:17:14 crc kubenswrapper[4873]: I1201 09:17:14.437447 4873 scope.go:117] "RemoveContainer" containerID="1a1db8f8327e8cf6dbb94b57ea898b250339bb198ae1ba595cb1d79805fa0db1" Dec 01 09:17:14 crc kubenswrapper[4873]: E1201 09:17:14.439434 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:17:16 crc kubenswrapper[4873]: I1201 09:17:16.710974 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-cmfkh"] Dec 01 09:17:16 crc kubenswrapper[4873]: E1201 09:17:16.712109 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d99932d2-5e5c-44ed-afa1-6f7616f5a03b" containerName="extract-content" Dec 01 09:17:16 crc kubenswrapper[4873]: I1201 09:17:16.712141 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="d99932d2-5e5c-44ed-afa1-6f7616f5a03b" containerName="extract-content" Dec 01 09:17:16 crc kubenswrapper[4873]: E1201 09:17:16.712162 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20057a0c-212d-4154-9556-dfbbf835c625" containerName="extract-content" Dec 01 09:17:16 crc kubenswrapper[4873]: I1201 09:17:16.712171 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="20057a0c-212d-4154-9556-dfbbf835c625" containerName="extract-content" Dec 01 09:17:16 crc kubenswrapper[4873]: E1201 09:17:16.712193 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20057a0c-212d-4154-9556-dfbbf835c625" containerName="extract-utilities" Dec 01 09:17:16 crc kubenswrapper[4873]: I1201 09:17:16.712202 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="20057a0c-212d-4154-9556-dfbbf835c625" containerName="extract-utilities" Dec 01 09:17:16 crc kubenswrapper[4873]: E1201 09:17:16.712235 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20057a0c-212d-4154-9556-dfbbf835c625" containerName="registry-server" Dec 01 09:17:16 crc kubenswrapper[4873]: I1201 09:17:16.712243 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="20057a0c-212d-4154-9556-dfbbf835c625" containerName="registry-server" Dec 01 09:17:16 crc kubenswrapper[4873]: E1201 09:17:16.712486 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d99932d2-5e5c-44ed-afa1-6f7616f5a03b" containerName="extract-utilities" Dec 01 09:17:16 crc kubenswrapper[4873]: I1201 09:17:16.712496 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="d99932d2-5e5c-44ed-afa1-6f7616f5a03b" containerName="extract-utilities" Dec 01 09:17:16 crc kubenswrapper[4873]: E1201 09:17:16.712514 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d99932d2-5e5c-44ed-afa1-6f7616f5a03b" containerName="registry-server" Dec 01 09:17:16 crc kubenswrapper[4873]: I1201 09:17:16.712522 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="d99932d2-5e5c-44ed-afa1-6f7616f5a03b" containerName="registry-server" Dec 01 09:17:16 crc kubenswrapper[4873]: I1201 09:17:16.715279 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="20057a0c-212d-4154-9556-dfbbf835c625" containerName="registry-server" Dec 01 09:17:16 crc kubenswrapper[4873]: I1201 09:17:16.715401 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="d99932d2-5e5c-44ed-afa1-6f7616f5a03b" containerName="registry-server" Dec 01 09:17:16 crc kubenswrapper[4873]: I1201 09:17:16.727038 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cmfkh"] Dec 01 09:17:16 crc kubenswrapper[4873]: I1201 09:17:16.727290 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cmfkh" Dec 01 09:17:16 crc kubenswrapper[4873]: I1201 09:17:16.787149 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08841d2a-0be3-4f17-b5f2-2635035626e8-utilities\") pod \"certified-operators-cmfkh\" (UID: \"08841d2a-0be3-4f17-b5f2-2635035626e8\") " pod="openshift-marketplace/certified-operators-cmfkh" Dec 01 09:17:16 crc kubenswrapper[4873]: I1201 09:17:16.787276 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08841d2a-0be3-4f17-b5f2-2635035626e8-catalog-content\") pod \"certified-operators-cmfkh\" (UID: \"08841d2a-0be3-4f17-b5f2-2635035626e8\") " pod="openshift-marketplace/certified-operators-cmfkh" Dec 01 09:17:16 crc kubenswrapper[4873]: I1201 09:17:16.787334 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdmnd\" (UniqueName: \"kubernetes.io/projected/08841d2a-0be3-4f17-b5f2-2635035626e8-kube-api-access-hdmnd\") pod \"certified-operators-cmfkh\" (UID: \"08841d2a-0be3-4f17-b5f2-2635035626e8\") " pod="openshift-marketplace/certified-operators-cmfkh" Dec 01 09:17:16 crc kubenswrapper[4873]: I1201 09:17:16.889525 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08841d2a-0be3-4f17-b5f2-2635035626e8-utilities\") pod \"certified-operators-cmfkh\" (UID: \"08841d2a-0be3-4f17-b5f2-2635035626e8\") " pod="openshift-marketplace/certified-operators-cmfkh" Dec 01 09:17:16 crc kubenswrapper[4873]: I1201 09:17:16.889631 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08841d2a-0be3-4f17-b5f2-2635035626e8-catalog-content\") pod \"certified-operators-cmfkh\" (UID: \"08841d2a-0be3-4f17-b5f2-2635035626e8\") " pod="openshift-marketplace/certified-operators-cmfkh" Dec 01 09:17:16 crc kubenswrapper[4873]: I1201 09:17:16.889681 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdmnd\" (UniqueName: \"kubernetes.io/projected/08841d2a-0be3-4f17-b5f2-2635035626e8-kube-api-access-hdmnd\") pod \"certified-operators-cmfkh\" (UID: \"08841d2a-0be3-4f17-b5f2-2635035626e8\") " pod="openshift-marketplace/certified-operators-cmfkh" Dec 01 09:17:16 crc kubenswrapper[4873]: I1201 09:17:16.890515 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08841d2a-0be3-4f17-b5f2-2635035626e8-utilities\") pod \"certified-operators-cmfkh\" (UID: \"08841d2a-0be3-4f17-b5f2-2635035626e8\") " pod="openshift-marketplace/certified-operators-cmfkh" Dec 01 09:17:16 crc kubenswrapper[4873]: I1201 09:17:16.890578 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08841d2a-0be3-4f17-b5f2-2635035626e8-catalog-content\") pod \"certified-operators-cmfkh\" (UID: \"08841d2a-0be3-4f17-b5f2-2635035626e8\") " pod="openshift-marketplace/certified-operators-cmfkh" Dec 01 09:17:16 crc kubenswrapper[4873]: I1201 09:17:16.915046 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdmnd\" (UniqueName: \"kubernetes.io/projected/08841d2a-0be3-4f17-b5f2-2635035626e8-kube-api-access-hdmnd\") pod \"certified-operators-cmfkh\" (UID: \"08841d2a-0be3-4f17-b5f2-2635035626e8\") " pod="openshift-marketplace/certified-operators-cmfkh" Dec 01 09:17:17 crc kubenswrapper[4873]: I1201 09:17:17.066481 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cmfkh" Dec 01 09:17:17 crc kubenswrapper[4873]: I1201 09:17:17.399745 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cmfkh"] Dec 01 09:17:17 crc kubenswrapper[4873]: I1201 09:17:17.493099 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cmfkh" event={"ID":"08841d2a-0be3-4f17-b5f2-2635035626e8","Type":"ContainerStarted","Data":"0631c55e3294fc8e68902a168fd10170c2607960cdcebeb5af156fab80925836"} Dec 01 09:17:18 crc kubenswrapper[4873]: I1201 09:17:18.505099 4873 generic.go:334] "Generic (PLEG): container finished" podID="08841d2a-0be3-4f17-b5f2-2635035626e8" containerID="2e2b54474abab467de3bb041aec8c926ca2989de14df5020ee7a0496d8491131" exitCode=0 Dec 01 09:17:18 crc kubenswrapper[4873]: I1201 09:17:18.505222 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cmfkh" event={"ID":"08841d2a-0be3-4f17-b5f2-2635035626e8","Type":"ContainerDied","Data":"2e2b54474abab467de3bb041aec8c926ca2989de14df5020ee7a0496d8491131"} Dec 01 09:17:20 crc kubenswrapper[4873]: I1201 09:17:20.534221 4873 generic.go:334] "Generic (PLEG): container finished" podID="08841d2a-0be3-4f17-b5f2-2635035626e8" containerID="1a7008798cc72d70f8bd03772f5d526babd16083b8b2efd7319d7b68fe06e93c" exitCode=0 Dec 01 09:17:20 crc kubenswrapper[4873]: I1201 09:17:20.534309 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cmfkh" event={"ID":"08841d2a-0be3-4f17-b5f2-2635035626e8","Type":"ContainerDied","Data":"1a7008798cc72d70f8bd03772f5d526babd16083b8b2efd7319d7b68fe06e93c"} Dec 01 09:17:23 crc kubenswrapper[4873]: I1201 09:17:23.566870 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cmfkh" event={"ID":"08841d2a-0be3-4f17-b5f2-2635035626e8","Type":"ContainerStarted","Data":"1b0c6cc7fa9e4441e621d907bca9c220c4a8a3b419d0d44b34fdfb80c029190c"} Dec 01 09:17:23 crc kubenswrapper[4873]: I1201 09:17:23.592824 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-cmfkh" podStartSLOduration=3.191453601 podStartE2EDuration="7.592800618s" podCreationTimestamp="2025-12-01 09:17:16 +0000 UTC" firstStartedPulling="2025-12-01 09:17:18.510663634 +0000 UTC m=+2214.412772173" lastFinishedPulling="2025-12-01 09:17:22.912010651 +0000 UTC m=+2218.814119190" observedRunningTime="2025-12-01 09:17:23.586684907 +0000 UTC m=+2219.488793446" watchObservedRunningTime="2025-12-01 09:17:23.592800618 +0000 UTC m=+2219.494909157" Dec 01 09:17:27 crc kubenswrapper[4873]: I1201 09:17:27.066718 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-cmfkh" Dec 01 09:17:27 crc kubenswrapper[4873]: I1201 09:17:27.067312 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-cmfkh" Dec 01 09:17:27 crc kubenswrapper[4873]: I1201 09:17:27.126180 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-cmfkh" Dec 01 09:17:27 crc kubenswrapper[4873]: I1201 09:17:27.421241 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-m7fx6"] Dec 01 09:17:27 crc kubenswrapper[4873]: I1201 09:17:27.424045 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m7fx6" Dec 01 09:17:27 crc kubenswrapper[4873]: I1201 09:17:27.430221 4873 scope.go:117] "RemoveContainer" containerID="1a1db8f8327e8cf6dbb94b57ea898b250339bb198ae1ba595cb1d79805fa0db1" Dec 01 09:17:27 crc kubenswrapper[4873]: E1201 09:17:27.430625 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:17:27 crc kubenswrapper[4873]: I1201 09:17:27.436481 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m7fx6"] Dec 01 09:17:27 crc kubenswrapper[4873]: I1201 09:17:27.551850 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db7ea12c-5eed-463b-84fc-2fede560560c-catalog-content\") pod \"community-operators-m7fx6\" (UID: \"db7ea12c-5eed-463b-84fc-2fede560560c\") " pod="openshift-marketplace/community-operators-m7fx6" Dec 01 09:17:27 crc kubenswrapper[4873]: I1201 09:17:27.551948 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwfwk\" (UniqueName: \"kubernetes.io/projected/db7ea12c-5eed-463b-84fc-2fede560560c-kube-api-access-hwfwk\") pod \"community-operators-m7fx6\" (UID: \"db7ea12c-5eed-463b-84fc-2fede560560c\") " pod="openshift-marketplace/community-operators-m7fx6" Dec 01 09:17:27 crc kubenswrapper[4873]: I1201 09:17:27.552184 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db7ea12c-5eed-463b-84fc-2fede560560c-utilities\") pod \"community-operators-m7fx6\" (UID: \"db7ea12c-5eed-463b-84fc-2fede560560c\") " pod="openshift-marketplace/community-operators-m7fx6" Dec 01 09:17:27 crc kubenswrapper[4873]: I1201 09:17:27.654300 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db7ea12c-5eed-463b-84fc-2fede560560c-utilities\") pod \"community-operators-m7fx6\" (UID: \"db7ea12c-5eed-463b-84fc-2fede560560c\") " pod="openshift-marketplace/community-operators-m7fx6" Dec 01 09:17:27 crc kubenswrapper[4873]: I1201 09:17:27.654440 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db7ea12c-5eed-463b-84fc-2fede560560c-catalog-content\") pod \"community-operators-m7fx6\" (UID: \"db7ea12c-5eed-463b-84fc-2fede560560c\") " pod="openshift-marketplace/community-operators-m7fx6" Dec 01 09:17:27 crc kubenswrapper[4873]: I1201 09:17:27.654488 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwfwk\" (UniqueName: \"kubernetes.io/projected/db7ea12c-5eed-463b-84fc-2fede560560c-kube-api-access-hwfwk\") pod \"community-operators-m7fx6\" (UID: \"db7ea12c-5eed-463b-84fc-2fede560560c\") " pod="openshift-marketplace/community-operators-m7fx6" Dec 01 09:17:27 crc kubenswrapper[4873]: I1201 09:17:27.654917 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db7ea12c-5eed-463b-84fc-2fede560560c-utilities\") pod \"community-operators-m7fx6\" (UID: \"db7ea12c-5eed-463b-84fc-2fede560560c\") " pod="openshift-marketplace/community-operators-m7fx6" Dec 01 09:17:27 crc kubenswrapper[4873]: I1201 09:17:27.654964 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db7ea12c-5eed-463b-84fc-2fede560560c-catalog-content\") pod \"community-operators-m7fx6\" (UID: \"db7ea12c-5eed-463b-84fc-2fede560560c\") " pod="openshift-marketplace/community-operators-m7fx6" Dec 01 09:17:27 crc kubenswrapper[4873]: I1201 09:17:27.683834 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwfwk\" (UniqueName: \"kubernetes.io/projected/db7ea12c-5eed-463b-84fc-2fede560560c-kube-api-access-hwfwk\") pod \"community-operators-m7fx6\" (UID: \"db7ea12c-5eed-463b-84fc-2fede560560c\") " pod="openshift-marketplace/community-operators-m7fx6" Dec 01 09:17:27 crc kubenswrapper[4873]: I1201 09:17:27.749110 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m7fx6" Dec 01 09:17:28 crc kubenswrapper[4873]: I1201 09:17:28.326239 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m7fx6"] Dec 01 09:17:28 crc kubenswrapper[4873]: I1201 09:17:28.619331 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m7fx6" event={"ID":"db7ea12c-5eed-463b-84fc-2fede560560c","Type":"ContainerStarted","Data":"5d000efc149782d89966bd34157798cd423a7c3f987f663966055e9bebccf63a"} Dec 01 09:17:29 crc kubenswrapper[4873]: I1201 09:17:29.631042 4873 generic.go:334] "Generic (PLEG): container finished" podID="db7ea12c-5eed-463b-84fc-2fede560560c" containerID="a49f7bf80665ccca7d107e14fbe609da7e444b834f8faa6459b3f5557b5be4aa" exitCode=0 Dec 01 09:17:29 crc kubenswrapper[4873]: I1201 09:17:29.631160 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m7fx6" event={"ID":"db7ea12c-5eed-463b-84fc-2fede560560c","Type":"ContainerDied","Data":"a49f7bf80665ccca7d107e14fbe609da7e444b834f8faa6459b3f5557b5be4aa"} Dec 01 09:17:31 crc kubenswrapper[4873]: I1201 09:17:31.653285 4873 generic.go:334] "Generic (PLEG): container finished" podID="db7ea12c-5eed-463b-84fc-2fede560560c" containerID="be0c50ffa0c48744956bd15420ef151587e66a45fc3d2ebf7dae4e59f893a865" exitCode=0 Dec 01 09:17:31 crc kubenswrapper[4873]: I1201 09:17:31.653800 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m7fx6" event={"ID":"db7ea12c-5eed-463b-84fc-2fede560560c","Type":"ContainerDied","Data":"be0c50ffa0c48744956bd15420ef151587e66a45fc3d2ebf7dae4e59f893a865"} Dec 01 09:17:32 crc kubenswrapper[4873]: I1201 09:17:32.666343 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m7fx6" event={"ID":"db7ea12c-5eed-463b-84fc-2fede560560c","Type":"ContainerStarted","Data":"bdc461a8abc7dd6eda2b61ed34b1952345773c54a6840b38ec8e6a4262965212"} Dec 01 09:17:32 crc kubenswrapper[4873]: I1201 09:17:32.694624 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-m7fx6" podStartSLOduration=3.140994704 podStartE2EDuration="5.69459619s" podCreationTimestamp="2025-12-01 09:17:27 +0000 UTC" firstStartedPulling="2025-12-01 09:17:29.634049316 +0000 UTC m=+2225.536157855" lastFinishedPulling="2025-12-01 09:17:32.187650802 +0000 UTC m=+2228.089759341" observedRunningTime="2025-12-01 09:17:32.693620826 +0000 UTC m=+2228.595729375" watchObservedRunningTime="2025-12-01 09:17:32.69459619 +0000 UTC m=+2228.596704729" Dec 01 09:17:37 crc kubenswrapper[4873]: I1201 09:17:37.144057 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-cmfkh" Dec 01 09:17:37 crc kubenswrapper[4873]: I1201 09:17:37.409533 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cmfkh"] Dec 01 09:17:37 crc kubenswrapper[4873]: I1201 09:17:37.715764 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-cmfkh" podUID="08841d2a-0be3-4f17-b5f2-2635035626e8" containerName="registry-server" containerID="cri-o://1b0c6cc7fa9e4441e621d907bca9c220c4a8a3b419d0d44b34fdfb80c029190c" gracePeriod=2 Dec 01 09:17:37 crc kubenswrapper[4873]: I1201 09:17:37.749333 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-m7fx6" Dec 01 09:17:37 crc kubenswrapper[4873]: I1201 09:17:37.750387 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-m7fx6" Dec 01 09:17:37 crc kubenswrapper[4873]: I1201 09:17:37.799922 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-m7fx6" Dec 01 09:17:38 crc kubenswrapper[4873]: I1201 09:17:38.183100 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cmfkh" Dec 01 09:17:38 crc kubenswrapper[4873]: I1201 09:17:38.347825 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08841d2a-0be3-4f17-b5f2-2635035626e8-utilities\") pod \"08841d2a-0be3-4f17-b5f2-2635035626e8\" (UID: \"08841d2a-0be3-4f17-b5f2-2635035626e8\") " Dec 01 09:17:38 crc kubenswrapper[4873]: I1201 09:17:38.348065 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdmnd\" (UniqueName: \"kubernetes.io/projected/08841d2a-0be3-4f17-b5f2-2635035626e8-kube-api-access-hdmnd\") pod \"08841d2a-0be3-4f17-b5f2-2635035626e8\" (UID: \"08841d2a-0be3-4f17-b5f2-2635035626e8\") " Dec 01 09:17:38 crc kubenswrapper[4873]: I1201 09:17:38.348366 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08841d2a-0be3-4f17-b5f2-2635035626e8-catalog-content\") pod \"08841d2a-0be3-4f17-b5f2-2635035626e8\" (UID: \"08841d2a-0be3-4f17-b5f2-2635035626e8\") " Dec 01 09:17:38 crc kubenswrapper[4873]: I1201 09:17:38.354738 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08841d2a-0be3-4f17-b5f2-2635035626e8-utilities" (OuterVolumeSpecName: "utilities") pod "08841d2a-0be3-4f17-b5f2-2635035626e8" (UID: "08841d2a-0be3-4f17-b5f2-2635035626e8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:17:38 crc kubenswrapper[4873]: I1201 09:17:38.364875 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08841d2a-0be3-4f17-b5f2-2635035626e8-kube-api-access-hdmnd" (OuterVolumeSpecName: "kube-api-access-hdmnd") pod "08841d2a-0be3-4f17-b5f2-2635035626e8" (UID: "08841d2a-0be3-4f17-b5f2-2635035626e8"). InnerVolumeSpecName "kube-api-access-hdmnd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:17:38 crc kubenswrapper[4873]: I1201 09:17:38.416973 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08841d2a-0be3-4f17-b5f2-2635035626e8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "08841d2a-0be3-4f17-b5f2-2635035626e8" (UID: "08841d2a-0be3-4f17-b5f2-2635035626e8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:17:38 crc kubenswrapper[4873]: I1201 09:17:38.456740 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdmnd\" (UniqueName: \"kubernetes.io/projected/08841d2a-0be3-4f17-b5f2-2635035626e8-kube-api-access-hdmnd\") on node \"crc\" DevicePath \"\"" Dec 01 09:17:38 crc kubenswrapper[4873]: I1201 09:17:38.457429 4873 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08841d2a-0be3-4f17-b5f2-2635035626e8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 09:17:38 crc kubenswrapper[4873]: I1201 09:17:38.457524 4873 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08841d2a-0be3-4f17-b5f2-2635035626e8-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 09:17:38 crc kubenswrapper[4873]: I1201 09:17:38.727874 4873 generic.go:334] "Generic (PLEG): container finished" podID="08841d2a-0be3-4f17-b5f2-2635035626e8" containerID="1b0c6cc7fa9e4441e621d907bca9c220c4a8a3b419d0d44b34fdfb80c029190c" exitCode=0 Dec 01 09:17:38 crc kubenswrapper[4873]: I1201 09:17:38.727964 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cmfkh" event={"ID":"08841d2a-0be3-4f17-b5f2-2635035626e8","Type":"ContainerDied","Data":"1b0c6cc7fa9e4441e621d907bca9c220c4a8a3b419d0d44b34fdfb80c029190c"} Dec 01 09:17:38 crc kubenswrapper[4873]: I1201 09:17:38.728055 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cmfkh" event={"ID":"08841d2a-0be3-4f17-b5f2-2635035626e8","Type":"ContainerDied","Data":"0631c55e3294fc8e68902a168fd10170c2607960cdcebeb5af156fab80925836"} Dec 01 09:17:38 crc kubenswrapper[4873]: I1201 09:17:38.728091 4873 scope.go:117] "RemoveContainer" containerID="1b0c6cc7fa9e4441e621d907bca9c220c4a8a3b419d0d44b34fdfb80c029190c" Dec 01 09:17:38 crc kubenswrapper[4873]: I1201 09:17:38.728364 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cmfkh" Dec 01 09:17:38 crc kubenswrapper[4873]: I1201 09:17:38.761569 4873 scope.go:117] "RemoveContainer" containerID="1a7008798cc72d70f8bd03772f5d526babd16083b8b2efd7319d7b68fe06e93c" Dec 01 09:17:38 crc kubenswrapper[4873]: I1201 09:17:38.769663 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cmfkh"] Dec 01 09:17:38 crc kubenswrapper[4873]: I1201 09:17:38.779395 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-cmfkh"] Dec 01 09:17:38 crc kubenswrapper[4873]: I1201 09:17:38.793741 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-m7fx6" Dec 01 09:17:38 crc kubenswrapper[4873]: I1201 09:17:38.825367 4873 scope.go:117] "RemoveContainer" containerID="2e2b54474abab467de3bb041aec8c926ca2989de14df5020ee7a0496d8491131" Dec 01 09:17:38 crc kubenswrapper[4873]: I1201 09:17:38.849439 4873 scope.go:117] "RemoveContainer" containerID="1b0c6cc7fa9e4441e621d907bca9c220c4a8a3b419d0d44b34fdfb80c029190c" Dec 01 09:17:38 crc kubenswrapper[4873]: E1201 09:17:38.850239 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b0c6cc7fa9e4441e621d907bca9c220c4a8a3b419d0d44b34fdfb80c029190c\": container with ID starting with 1b0c6cc7fa9e4441e621d907bca9c220c4a8a3b419d0d44b34fdfb80c029190c not found: ID does not exist" containerID="1b0c6cc7fa9e4441e621d907bca9c220c4a8a3b419d0d44b34fdfb80c029190c" Dec 01 09:17:38 crc kubenswrapper[4873]: I1201 09:17:38.850281 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b0c6cc7fa9e4441e621d907bca9c220c4a8a3b419d0d44b34fdfb80c029190c"} err="failed to get container status \"1b0c6cc7fa9e4441e621d907bca9c220c4a8a3b419d0d44b34fdfb80c029190c\": rpc error: code = NotFound desc = could not find container \"1b0c6cc7fa9e4441e621d907bca9c220c4a8a3b419d0d44b34fdfb80c029190c\": container with ID starting with 1b0c6cc7fa9e4441e621d907bca9c220c4a8a3b419d0d44b34fdfb80c029190c not found: ID does not exist" Dec 01 09:17:38 crc kubenswrapper[4873]: I1201 09:17:38.850313 4873 scope.go:117] "RemoveContainer" containerID="1a7008798cc72d70f8bd03772f5d526babd16083b8b2efd7319d7b68fe06e93c" Dec 01 09:17:38 crc kubenswrapper[4873]: E1201 09:17:38.850842 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a7008798cc72d70f8bd03772f5d526babd16083b8b2efd7319d7b68fe06e93c\": container with ID starting with 1a7008798cc72d70f8bd03772f5d526babd16083b8b2efd7319d7b68fe06e93c not found: ID does not exist" containerID="1a7008798cc72d70f8bd03772f5d526babd16083b8b2efd7319d7b68fe06e93c" Dec 01 09:17:38 crc kubenswrapper[4873]: I1201 09:17:38.850876 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a7008798cc72d70f8bd03772f5d526babd16083b8b2efd7319d7b68fe06e93c"} err="failed to get container status \"1a7008798cc72d70f8bd03772f5d526babd16083b8b2efd7319d7b68fe06e93c\": rpc error: code = NotFound desc = could not find container \"1a7008798cc72d70f8bd03772f5d526babd16083b8b2efd7319d7b68fe06e93c\": container with ID starting with 1a7008798cc72d70f8bd03772f5d526babd16083b8b2efd7319d7b68fe06e93c not found: ID does not exist" Dec 01 09:17:38 crc kubenswrapper[4873]: I1201 09:17:38.850895 4873 scope.go:117] "RemoveContainer" containerID="2e2b54474abab467de3bb041aec8c926ca2989de14df5020ee7a0496d8491131" Dec 01 09:17:38 crc kubenswrapper[4873]: E1201 09:17:38.851368 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e2b54474abab467de3bb041aec8c926ca2989de14df5020ee7a0496d8491131\": container with ID starting with 2e2b54474abab467de3bb041aec8c926ca2989de14df5020ee7a0496d8491131 not found: ID does not exist" containerID="2e2b54474abab467de3bb041aec8c926ca2989de14df5020ee7a0496d8491131" Dec 01 09:17:38 crc kubenswrapper[4873]: I1201 09:17:38.851459 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e2b54474abab467de3bb041aec8c926ca2989de14df5020ee7a0496d8491131"} err="failed to get container status \"2e2b54474abab467de3bb041aec8c926ca2989de14df5020ee7a0496d8491131\": rpc error: code = NotFound desc = could not find container \"2e2b54474abab467de3bb041aec8c926ca2989de14df5020ee7a0496d8491131\": container with ID starting with 2e2b54474abab467de3bb041aec8c926ca2989de14df5020ee7a0496d8491131 not found: ID does not exist" Dec 01 09:17:40 crc kubenswrapper[4873]: I1201 09:17:40.210785 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-m7fx6"] Dec 01 09:17:40 crc kubenswrapper[4873]: I1201 09:17:40.443334 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08841d2a-0be3-4f17-b5f2-2635035626e8" path="/var/lib/kubelet/pods/08841d2a-0be3-4f17-b5f2-2635035626e8/volumes" Dec 01 09:17:41 crc kubenswrapper[4873]: I1201 09:17:41.430456 4873 scope.go:117] "RemoveContainer" containerID="1a1db8f8327e8cf6dbb94b57ea898b250339bb198ae1ba595cb1d79805fa0db1" Dec 01 09:17:41 crc kubenswrapper[4873]: E1201 09:17:41.430799 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:17:41 crc kubenswrapper[4873]: I1201 09:17:41.785394 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-m7fx6" podUID="db7ea12c-5eed-463b-84fc-2fede560560c" containerName="registry-server" containerID="cri-o://bdc461a8abc7dd6eda2b61ed34b1952345773c54a6840b38ec8e6a4262965212" gracePeriod=2 Dec 01 09:17:42 crc kubenswrapper[4873]: I1201 09:17:42.306560 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m7fx6" Dec 01 09:17:42 crc kubenswrapper[4873]: I1201 09:17:42.457250 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db7ea12c-5eed-463b-84fc-2fede560560c-utilities\") pod \"db7ea12c-5eed-463b-84fc-2fede560560c\" (UID: \"db7ea12c-5eed-463b-84fc-2fede560560c\") " Dec 01 09:17:42 crc kubenswrapper[4873]: I1201 09:17:42.457849 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hwfwk\" (UniqueName: \"kubernetes.io/projected/db7ea12c-5eed-463b-84fc-2fede560560c-kube-api-access-hwfwk\") pod \"db7ea12c-5eed-463b-84fc-2fede560560c\" (UID: \"db7ea12c-5eed-463b-84fc-2fede560560c\") " Dec 01 09:17:42 crc kubenswrapper[4873]: I1201 09:17:42.458005 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db7ea12c-5eed-463b-84fc-2fede560560c-catalog-content\") pod \"db7ea12c-5eed-463b-84fc-2fede560560c\" (UID: \"db7ea12c-5eed-463b-84fc-2fede560560c\") " Dec 01 09:17:42 crc kubenswrapper[4873]: I1201 09:17:42.467766 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db7ea12c-5eed-463b-84fc-2fede560560c-utilities" (OuterVolumeSpecName: "utilities") pod "db7ea12c-5eed-463b-84fc-2fede560560c" (UID: "db7ea12c-5eed-463b-84fc-2fede560560c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:17:42 crc kubenswrapper[4873]: I1201 09:17:42.469727 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db7ea12c-5eed-463b-84fc-2fede560560c-kube-api-access-hwfwk" (OuterVolumeSpecName: "kube-api-access-hwfwk") pod "db7ea12c-5eed-463b-84fc-2fede560560c" (UID: "db7ea12c-5eed-463b-84fc-2fede560560c"). InnerVolumeSpecName "kube-api-access-hwfwk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:17:42 crc kubenswrapper[4873]: I1201 09:17:42.513747 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db7ea12c-5eed-463b-84fc-2fede560560c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "db7ea12c-5eed-463b-84fc-2fede560560c" (UID: "db7ea12c-5eed-463b-84fc-2fede560560c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:17:42 crc kubenswrapper[4873]: I1201 09:17:42.561251 4873 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db7ea12c-5eed-463b-84fc-2fede560560c-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 09:17:42 crc kubenswrapper[4873]: I1201 09:17:42.561305 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hwfwk\" (UniqueName: \"kubernetes.io/projected/db7ea12c-5eed-463b-84fc-2fede560560c-kube-api-access-hwfwk\") on node \"crc\" DevicePath \"\"" Dec 01 09:17:42 crc kubenswrapper[4873]: I1201 09:17:42.561318 4873 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db7ea12c-5eed-463b-84fc-2fede560560c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 09:17:42 crc kubenswrapper[4873]: I1201 09:17:42.797091 4873 generic.go:334] "Generic (PLEG): container finished" podID="db7ea12c-5eed-463b-84fc-2fede560560c" containerID="bdc461a8abc7dd6eda2b61ed34b1952345773c54a6840b38ec8e6a4262965212" exitCode=0 Dec 01 09:17:42 crc kubenswrapper[4873]: I1201 09:17:42.797156 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m7fx6" event={"ID":"db7ea12c-5eed-463b-84fc-2fede560560c","Type":"ContainerDied","Data":"bdc461a8abc7dd6eda2b61ed34b1952345773c54a6840b38ec8e6a4262965212"} Dec 01 09:17:42 crc kubenswrapper[4873]: I1201 09:17:42.797196 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m7fx6" event={"ID":"db7ea12c-5eed-463b-84fc-2fede560560c","Type":"ContainerDied","Data":"5d000efc149782d89966bd34157798cd423a7c3f987f663966055e9bebccf63a"} Dec 01 09:17:42 crc kubenswrapper[4873]: I1201 09:17:42.797224 4873 scope.go:117] "RemoveContainer" containerID="bdc461a8abc7dd6eda2b61ed34b1952345773c54a6840b38ec8e6a4262965212" Dec 01 09:17:42 crc kubenswrapper[4873]: I1201 09:17:42.797253 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m7fx6" Dec 01 09:17:42 crc kubenswrapper[4873]: I1201 09:17:42.821555 4873 scope.go:117] "RemoveContainer" containerID="be0c50ffa0c48744956bd15420ef151587e66a45fc3d2ebf7dae4e59f893a865" Dec 01 09:17:42 crc kubenswrapper[4873]: I1201 09:17:42.840133 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-m7fx6"] Dec 01 09:17:42 crc kubenswrapper[4873]: I1201 09:17:42.847807 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-m7fx6"] Dec 01 09:17:42 crc kubenswrapper[4873]: I1201 09:17:42.866585 4873 scope.go:117] "RemoveContainer" containerID="a49f7bf80665ccca7d107e14fbe609da7e444b834f8faa6459b3f5557b5be4aa" Dec 01 09:17:42 crc kubenswrapper[4873]: I1201 09:17:42.905849 4873 scope.go:117] "RemoveContainer" containerID="bdc461a8abc7dd6eda2b61ed34b1952345773c54a6840b38ec8e6a4262965212" Dec 01 09:17:42 crc kubenswrapper[4873]: E1201 09:17:42.906469 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bdc461a8abc7dd6eda2b61ed34b1952345773c54a6840b38ec8e6a4262965212\": container with ID starting with bdc461a8abc7dd6eda2b61ed34b1952345773c54a6840b38ec8e6a4262965212 not found: ID does not exist" containerID="bdc461a8abc7dd6eda2b61ed34b1952345773c54a6840b38ec8e6a4262965212" Dec 01 09:17:42 crc kubenswrapper[4873]: I1201 09:17:42.906520 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdc461a8abc7dd6eda2b61ed34b1952345773c54a6840b38ec8e6a4262965212"} err="failed to get container status \"bdc461a8abc7dd6eda2b61ed34b1952345773c54a6840b38ec8e6a4262965212\": rpc error: code = NotFound desc = could not find container \"bdc461a8abc7dd6eda2b61ed34b1952345773c54a6840b38ec8e6a4262965212\": container with ID starting with bdc461a8abc7dd6eda2b61ed34b1952345773c54a6840b38ec8e6a4262965212 not found: ID does not exist" Dec 01 09:17:42 crc kubenswrapper[4873]: I1201 09:17:42.906556 4873 scope.go:117] "RemoveContainer" containerID="be0c50ffa0c48744956bd15420ef151587e66a45fc3d2ebf7dae4e59f893a865" Dec 01 09:17:42 crc kubenswrapper[4873]: E1201 09:17:42.907321 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be0c50ffa0c48744956bd15420ef151587e66a45fc3d2ebf7dae4e59f893a865\": container with ID starting with be0c50ffa0c48744956bd15420ef151587e66a45fc3d2ebf7dae4e59f893a865 not found: ID does not exist" containerID="be0c50ffa0c48744956bd15420ef151587e66a45fc3d2ebf7dae4e59f893a865" Dec 01 09:17:42 crc kubenswrapper[4873]: I1201 09:17:42.907346 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be0c50ffa0c48744956bd15420ef151587e66a45fc3d2ebf7dae4e59f893a865"} err="failed to get container status \"be0c50ffa0c48744956bd15420ef151587e66a45fc3d2ebf7dae4e59f893a865\": rpc error: code = NotFound desc = could not find container \"be0c50ffa0c48744956bd15420ef151587e66a45fc3d2ebf7dae4e59f893a865\": container with ID starting with be0c50ffa0c48744956bd15420ef151587e66a45fc3d2ebf7dae4e59f893a865 not found: ID does not exist" Dec 01 09:17:42 crc kubenswrapper[4873]: I1201 09:17:42.907361 4873 scope.go:117] "RemoveContainer" containerID="a49f7bf80665ccca7d107e14fbe609da7e444b834f8faa6459b3f5557b5be4aa" Dec 01 09:17:42 crc kubenswrapper[4873]: E1201 09:17:42.907622 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a49f7bf80665ccca7d107e14fbe609da7e444b834f8faa6459b3f5557b5be4aa\": container with ID starting with a49f7bf80665ccca7d107e14fbe609da7e444b834f8faa6459b3f5557b5be4aa not found: ID does not exist" containerID="a49f7bf80665ccca7d107e14fbe609da7e444b834f8faa6459b3f5557b5be4aa" Dec 01 09:17:42 crc kubenswrapper[4873]: I1201 09:17:42.907640 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a49f7bf80665ccca7d107e14fbe609da7e444b834f8faa6459b3f5557b5be4aa"} err="failed to get container status \"a49f7bf80665ccca7d107e14fbe609da7e444b834f8faa6459b3f5557b5be4aa\": rpc error: code = NotFound desc = could not find container \"a49f7bf80665ccca7d107e14fbe609da7e444b834f8faa6459b3f5557b5be4aa\": container with ID starting with a49f7bf80665ccca7d107e14fbe609da7e444b834f8faa6459b3f5557b5be4aa not found: ID does not exist" Dec 01 09:17:44 crc kubenswrapper[4873]: I1201 09:17:44.445215 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db7ea12c-5eed-463b-84fc-2fede560560c" path="/var/lib/kubelet/pods/db7ea12c-5eed-463b-84fc-2fede560560c/volumes" Dec 01 09:17:53 crc kubenswrapper[4873]: I1201 09:17:53.430430 4873 scope.go:117] "RemoveContainer" containerID="1a1db8f8327e8cf6dbb94b57ea898b250339bb198ae1ba595cb1d79805fa0db1" Dec 01 09:17:53 crc kubenswrapper[4873]: E1201 09:17:53.431613 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:17:58 crc kubenswrapper[4873]: I1201 09:17:58.480400 4873 scope.go:117] "RemoveContainer" containerID="e3f8450fc3731b547de2770c41b124219ae667f8edbe702dec2cda9d4e34f94f" Dec 01 09:17:58 crc kubenswrapper[4873]: I1201 09:17:58.526572 4873 scope.go:117] "RemoveContainer" containerID="f62011a713228303254b933fa9fb59f221386364ba14686a1da3aa9692dd3eda" Dec 01 09:18:06 crc kubenswrapper[4873]: I1201 09:18:06.430342 4873 scope.go:117] "RemoveContainer" containerID="1a1db8f8327e8cf6dbb94b57ea898b250339bb198ae1ba595cb1d79805fa0db1" Dec 01 09:18:06 crc kubenswrapper[4873]: E1201 09:18:06.431741 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:18:07 crc kubenswrapper[4873]: I1201 09:18:07.043349 4873 generic.go:334] "Generic (PLEG): container finished" podID="cd2a1185-cebc-4f89-88c2-63cfebe97759" containerID="24c1eb182b53b52b56b83093b868d7e8915318fe5cc8af06f81effeff6d5dca1" exitCode=0 Dec 01 09:18:07 crc kubenswrapper[4873]: I1201 09:18:07.043441 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wfccb" event={"ID":"cd2a1185-cebc-4f89-88c2-63cfebe97759","Type":"ContainerDied","Data":"24c1eb182b53b52b56b83093b868d7e8915318fe5cc8af06f81effeff6d5dca1"} Dec 01 09:18:08 crc kubenswrapper[4873]: I1201 09:18:08.475481 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wfccb" Dec 01 09:18:08 crc kubenswrapper[4873]: I1201 09:18:08.668924 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd2a1185-cebc-4f89-88c2-63cfebe97759-bootstrap-combined-ca-bundle\") pod \"cd2a1185-cebc-4f89-88c2-63cfebe97759\" (UID: \"cd2a1185-cebc-4f89-88c2-63cfebe97759\") " Dec 01 09:18:08 crc kubenswrapper[4873]: I1201 09:18:08.669068 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cd2a1185-cebc-4f89-88c2-63cfebe97759-inventory\") pod \"cd2a1185-cebc-4f89-88c2-63cfebe97759\" (UID: \"cd2a1185-cebc-4f89-88c2-63cfebe97759\") " Dec 01 09:18:08 crc kubenswrapper[4873]: I1201 09:18:08.669139 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cd2a1185-cebc-4f89-88c2-63cfebe97759-ceph\") pod \"cd2a1185-cebc-4f89-88c2-63cfebe97759\" (UID: \"cd2a1185-cebc-4f89-88c2-63cfebe97759\") " Dec 01 09:18:08 crc kubenswrapper[4873]: I1201 09:18:08.669237 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cd2a1185-cebc-4f89-88c2-63cfebe97759-ssh-key\") pod \"cd2a1185-cebc-4f89-88c2-63cfebe97759\" (UID: \"cd2a1185-cebc-4f89-88c2-63cfebe97759\") " Dec 01 09:18:08 crc kubenswrapper[4873]: I1201 09:18:08.669364 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89gv8\" (UniqueName: \"kubernetes.io/projected/cd2a1185-cebc-4f89-88c2-63cfebe97759-kube-api-access-89gv8\") pod \"cd2a1185-cebc-4f89-88c2-63cfebe97759\" (UID: \"cd2a1185-cebc-4f89-88c2-63cfebe97759\") " Dec 01 09:18:08 crc kubenswrapper[4873]: I1201 09:18:08.676574 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd2a1185-cebc-4f89-88c2-63cfebe97759-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "cd2a1185-cebc-4f89-88c2-63cfebe97759" (UID: "cd2a1185-cebc-4f89-88c2-63cfebe97759"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:18:08 crc kubenswrapper[4873]: I1201 09:18:08.678433 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd2a1185-cebc-4f89-88c2-63cfebe97759-kube-api-access-89gv8" (OuterVolumeSpecName: "kube-api-access-89gv8") pod "cd2a1185-cebc-4f89-88c2-63cfebe97759" (UID: "cd2a1185-cebc-4f89-88c2-63cfebe97759"). InnerVolumeSpecName "kube-api-access-89gv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:18:08 crc kubenswrapper[4873]: I1201 09:18:08.679720 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd2a1185-cebc-4f89-88c2-63cfebe97759-ceph" (OuterVolumeSpecName: "ceph") pod "cd2a1185-cebc-4f89-88c2-63cfebe97759" (UID: "cd2a1185-cebc-4f89-88c2-63cfebe97759"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:18:08 crc kubenswrapper[4873]: I1201 09:18:08.700425 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd2a1185-cebc-4f89-88c2-63cfebe97759-inventory" (OuterVolumeSpecName: "inventory") pod "cd2a1185-cebc-4f89-88c2-63cfebe97759" (UID: "cd2a1185-cebc-4f89-88c2-63cfebe97759"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:18:08 crc kubenswrapper[4873]: I1201 09:18:08.708813 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd2a1185-cebc-4f89-88c2-63cfebe97759-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "cd2a1185-cebc-4f89-88c2-63cfebe97759" (UID: "cd2a1185-cebc-4f89-88c2-63cfebe97759"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:18:08 crc kubenswrapper[4873]: I1201 09:18:08.772048 4873 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cd2a1185-cebc-4f89-88c2-63cfebe97759-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 09:18:08 crc kubenswrapper[4873]: I1201 09:18:08.772100 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89gv8\" (UniqueName: \"kubernetes.io/projected/cd2a1185-cebc-4f89-88c2-63cfebe97759-kube-api-access-89gv8\") on node \"crc\" DevicePath \"\"" Dec 01 09:18:08 crc kubenswrapper[4873]: I1201 09:18:08.772114 4873 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd2a1185-cebc-4f89-88c2-63cfebe97759-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:18:08 crc kubenswrapper[4873]: I1201 09:18:08.772124 4873 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cd2a1185-cebc-4f89-88c2-63cfebe97759-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 09:18:08 crc kubenswrapper[4873]: I1201 09:18:08.772136 4873 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cd2a1185-cebc-4f89-88c2-63cfebe97759-ceph\") on node \"crc\" DevicePath \"\"" Dec 01 09:18:09 crc kubenswrapper[4873]: I1201 09:18:09.064667 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wfccb" event={"ID":"cd2a1185-cebc-4f89-88c2-63cfebe97759","Type":"ContainerDied","Data":"7117128aaf609a98be60540f9d69bfdff4358feade008d8db8588c0e5391bfb1"} Dec 01 09:18:09 crc kubenswrapper[4873]: I1201 09:18:09.064726 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7117128aaf609a98be60540f9d69bfdff4358feade008d8db8588c0e5391bfb1" Dec 01 09:18:09 crc kubenswrapper[4873]: I1201 09:18:09.065226 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wfccb" Dec 01 09:18:09 crc kubenswrapper[4873]: I1201 09:18:09.187163 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-lcck8"] Dec 01 09:18:09 crc kubenswrapper[4873]: E1201 09:18:09.187645 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08841d2a-0be3-4f17-b5f2-2635035626e8" containerName="extract-utilities" Dec 01 09:18:09 crc kubenswrapper[4873]: I1201 09:18:09.187667 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="08841d2a-0be3-4f17-b5f2-2635035626e8" containerName="extract-utilities" Dec 01 09:18:09 crc kubenswrapper[4873]: E1201 09:18:09.187688 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db7ea12c-5eed-463b-84fc-2fede560560c" containerName="extract-utilities" Dec 01 09:18:09 crc kubenswrapper[4873]: I1201 09:18:09.187697 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="db7ea12c-5eed-463b-84fc-2fede560560c" containerName="extract-utilities" Dec 01 09:18:09 crc kubenswrapper[4873]: E1201 09:18:09.187718 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db7ea12c-5eed-463b-84fc-2fede560560c" containerName="extract-content" Dec 01 09:18:09 crc kubenswrapper[4873]: I1201 09:18:09.187727 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="db7ea12c-5eed-463b-84fc-2fede560560c" containerName="extract-content" Dec 01 09:18:09 crc kubenswrapper[4873]: E1201 09:18:09.187737 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08841d2a-0be3-4f17-b5f2-2635035626e8" containerName="extract-content" Dec 01 09:18:09 crc kubenswrapper[4873]: I1201 09:18:09.187747 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="08841d2a-0be3-4f17-b5f2-2635035626e8" containerName="extract-content" Dec 01 09:18:09 crc kubenswrapper[4873]: E1201 09:18:09.187782 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08841d2a-0be3-4f17-b5f2-2635035626e8" containerName="registry-server" Dec 01 09:18:09 crc kubenswrapper[4873]: I1201 09:18:09.187792 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="08841d2a-0be3-4f17-b5f2-2635035626e8" containerName="registry-server" Dec 01 09:18:09 crc kubenswrapper[4873]: E1201 09:18:09.187817 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db7ea12c-5eed-463b-84fc-2fede560560c" containerName="registry-server" Dec 01 09:18:09 crc kubenswrapper[4873]: I1201 09:18:09.187823 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="db7ea12c-5eed-463b-84fc-2fede560560c" containerName="registry-server" Dec 01 09:18:09 crc kubenswrapper[4873]: E1201 09:18:09.187832 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd2a1185-cebc-4f89-88c2-63cfebe97759" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 01 09:18:09 crc kubenswrapper[4873]: I1201 09:18:09.187839 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd2a1185-cebc-4f89-88c2-63cfebe97759" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 01 09:18:09 crc kubenswrapper[4873]: I1201 09:18:09.188048 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="08841d2a-0be3-4f17-b5f2-2635035626e8" containerName="registry-server" Dec 01 09:18:09 crc kubenswrapper[4873]: I1201 09:18:09.188065 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd2a1185-cebc-4f89-88c2-63cfebe97759" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 01 09:18:09 crc kubenswrapper[4873]: I1201 09:18:09.188080 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="db7ea12c-5eed-463b-84fc-2fede560560c" containerName="registry-server" Dec 01 09:18:09 crc kubenswrapper[4873]: I1201 09:18:09.188789 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-lcck8" Dec 01 09:18:09 crc kubenswrapper[4873]: I1201 09:18:09.198715 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 09:18:09 crc kubenswrapper[4873]: I1201 09:18:09.199411 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 01 09:18:09 crc kubenswrapper[4873]: I1201 09:18:09.199563 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 09:18:09 crc kubenswrapper[4873]: I1201 09:18:09.199810 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9ghqw" Dec 01 09:18:09 crc kubenswrapper[4873]: I1201 09:18:09.199971 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 09:18:09 crc kubenswrapper[4873]: I1201 09:18:09.204210 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-lcck8"] Dec 01 09:18:09 crc kubenswrapper[4873]: I1201 09:18:09.386861 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zb97c\" (UniqueName: \"kubernetes.io/projected/99a70cc5-d0e6-4066-9fdb-7524bf04b7ab-kube-api-access-zb97c\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-lcck8\" (UID: \"99a70cc5-d0e6-4066-9fdb-7524bf04b7ab\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-lcck8" Dec 01 09:18:09 crc kubenswrapper[4873]: I1201 09:18:09.387054 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/99a70cc5-d0e6-4066-9fdb-7524bf04b7ab-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-lcck8\" (UID: \"99a70cc5-d0e6-4066-9fdb-7524bf04b7ab\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-lcck8" Dec 01 09:18:09 crc kubenswrapper[4873]: I1201 09:18:09.387181 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/99a70cc5-d0e6-4066-9fdb-7524bf04b7ab-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-lcck8\" (UID: \"99a70cc5-d0e6-4066-9fdb-7524bf04b7ab\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-lcck8" Dec 01 09:18:09 crc kubenswrapper[4873]: I1201 09:18:09.387338 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/99a70cc5-d0e6-4066-9fdb-7524bf04b7ab-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-lcck8\" (UID: \"99a70cc5-d0e6-4066-9fdb-7524bf04b7ab\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-lcck8" Dec 01 09:18:09 crc kubenswrapper[4873]: I1201 09:18:09.489441 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/99a70cc5-d0e6-4066-9fdb-7524bf04b7ab-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-lcck8\" (UID: \"99a70cc5-d0e6-4066-9fdb-7524bf04b7ab\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-lcck8" Dec 01 09:18:09 crc kubenswrapper[4873]: I1201 09:18:09.489500 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/99a70cc5-d0e6-4066-9fdb-7524bf04b7ab-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-lcck8\" (UID: \"99a70cc5-d0e6-4066-9fdb-7524bf04b7ab\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-lcck8" Dec 01 09:18:09 crc kubenswrapper[4873]: I1201 09:18:09.489556 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/99a70cc5-d0e6-4066-9fdb-7524bf04b7ab-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-lcck8\" (UID: \"99a70cc5-d0e6-4066-9fdb-7524bf04b7ab\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-lcck8" Dec 01 09:18:09 crc kubenswrapper[4873]: I1201 09:18:09.489668 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zb97c\" (UniqueName: \"kubernetes.io/projected/99a70cc5-d0e6-4066-9fdb-7524bf04b7ab-kube-api-access-zb97c\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-lcck8\" (UID: \"99a70cc5-d0e6-4066-9fdb-7524bf04b7ab\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-lcck8" Dec 01 09:18:09 crc kubenswrapper[4873]: I1201 09:18:09.493779 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/99a70cc5-d0e6-4066-9fdb-7524bf04b7ab-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-lcck8\" (UID: \"99a70cc5-d0e6-4066-9fdb-7524bf04b7ab\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-lcck8" Dec 01 09:18:09 crc kubenswrapper[4873]: I1201 09:18:09.493803 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/99a70cc5-d0e6-4066-9fdb-7524bf04b7ab-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-lcck8\" (UID: \"99a70cc5-d0e6-4066-9fdb-7524bf04b7ab\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-lcck8" Dec 01 09:18:09 crc kubenswrapper[4873]: I1201 09:18:09.495497 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/99a70cc5-d0e6-4066-9fdb-7524bf04b7ab-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-lcck8\" (UID: \"99a70cc5-d0e6-4066-9fdb-7524bf04b7ab\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-lcck8" Dec 01 09:18:09 crc kubenswrapper[4873]: I1201 09:18:09.545266 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zb97c\" (UniqueName: \"kubernetes.io/projected/99a70cc5-d0e6-4066-9fdb-7524bf04b7ab-kube-api-access-zb97c\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-lcck8\" (UID: \"99a70cc5-d0e6-4066-9fdb-7524bf04b7ab\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-lcck8" Dec 01 09:18:09 crc kubenswrapper[4873]: I1201 09:18:09.813369 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-lcck8" Dec 01 09:18:10 crc kubenswrapper[4873]: I1201 09:18:10.379197 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-lcck8"] Dec 01 09:18:11 crc kubenswrapper[4873]: I1201 09:18:11.113303 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-lcck8" event={"ID":"99a70cc5-d0e6-4066-9fdb-7524bf04b7ab","Type":"ContainerStarted","Data":"3223ddd4dc8bbb4e79a040ecafc989178550c5c275d093617ede481e4c32fd82"} Dec 01 09:18:12 crc kubenswrapper[4873]: I1201 09:18:12.124124 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-lcck8" event={"ID":"99a70cc5-d0e6-4066-9fdb-7524bf04b7ab","Type":"ContainerStarted","Data":"5bf87b7c6355a4d2e1e9792a81b5ffa6bfda228f98bbb61b12c8735bd478496f"} Dec 01 09:18:18 crc kubenswrapper[4873]: I1201 09:18:18.430814 4873 scope.go:117] "RemoveContainer" containerID="1a1db8f8327e8cf6dbb94b57ea898b250339bb198ae1ba595cb1d79805fa0db1" Dec 01 09:18:18 crc kubenswrapper[4873]: E1201 09:18:18.431686 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:18:30 crc kubenswrapper[4873]: I1201 09:18:30.430792 4873 scope.go:117] "RemoveContainer" containerID="1a1db8f8327e8cf6dbb94b57ea898b250339bb198ae1ba595cb1d79805fa0db1" Dec 01 09:18:30 crc kubenswrapper[4873]: E1201 09:18:30.431661 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:18:36 crc kubenswrapper[4873]: I1201 09:18:36.353962 4873 generic.go:334] "Generic (PLEG): container finished" podID="99a70cc5-d0e6-4066-9fdb-7524bf04b7ab" containerID="5bf87b7c6355a4d2e1e9792a81b5ffa6bfda228f98bbb61b12c8735bd478496f" exitCode=0 Dec 01 09:18:36 crc kubenswrapper[4873]: I1201 09:18:36.354138 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-lcck8" event={"ID":"99a70cc5-d0e6-4066-9fdb-7524bf04b7ab","Type":"ContainerDied","Data":"5bf87b7c6355a4d2e1e9792a81b5ffa6bfda228f98bbb61b12c8735bd478496f"} Dec 01 09:18:37 crc kubenswrapper[4873]: I1201 09:18:37.828249 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-lcck8" Dec 01 09:18:38 crc kubenswrapper[4873]: I1201 09:18:38.025316 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zb97c\" (UniqueName: \"kubernetes.io/projected/99a70cc5-d0e6-4066-9fdb-7524bf04b7ab-kube-api-access-zb97c\") pod \"99a70cc5-d0e6-4066-9fdb-7524bf04b7ab\" (UID: \"99a70cc5-d0e6-4066-9fdb-7524bf04b7ab\") " Dec 01 09:18:38 crc kubenswrapper[4873]: I1201 09:18:38.025409 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/99a70cc5-d0e6-4066-9fdb-7524bf04b7ab-ssh-key\") pod \"99a70cc5-d0e6-4066-9fdb-7524bf04b7ab\" (UID: \"99a70cc5-d0e6-4066-9fdb-7524bf04b7ab\") " Dec 01 09:18:38 crc kubenswrapper[4873]: I1201 09:18:38.025522 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/99a70cc5-d0e6-4066-9fdb-7524bf04b7ab-inventory\") pod \"99a70cc5-d0e6-4066-9fdb-7524bf04b7ab\" (UID: \"99a70cc5-d0e6-4066-9fdb-7524bf04b7ab\") " Dec 01 09:18:38 crc kubenswrapper[4873]: I1201 09:18:38.025560 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/99a70cc5-d0e6-4066-9fdb-7524bf04b7ab-ceph\") pod \"99a70cc5-d0e6-4066-9fdb-7524bf04b7ab\" (UID: \"99a70cc5-d0e6-4066-9fdb-7524bf04b7ab\") " Dec 01 09:18:38 crc kubenswrapper[4873]: I1201 09:18:38.035349 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99a70cc5-d0e6-4066-9fdb-7524bf04b7ab-kube-api-access-zb97c" (OuterVolumeSpecName: "kube-api-access-zb97c") pod "99a70cc5-d0e6-4066-9fdb-7524bf04b7ab" (UID: "99a70cc5-d0e6-4066-9fdb-7524bf04b7ab"). InnerVolumeSpecName "kube-api-access-zb97c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:18:38 crc kubenswrapper[4873]: I1201 09:18:38.035554 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99a70cc5-d0e6-4066-9fdb-7524bf04b7ab-ceph" (OuterVolumeSpecName: "ceph") pod "99a70cc5-d0e6-4066-9fdb-7524bf04b7ab" (UID: "99a70cc5-d0e6-4066-9fdb-7524bf04b7ab"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:18:38 crc kubenswrapper[4873]: I1201 09:18:38.063796 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99a70cc5-d0e6-4066-9fdb-7524bf04b7ab-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "99a70cc5-d0e6-4066-9fdb-7524bf04b7ab" (UID: "99a70cc5-d0e6-4066-9fdb-7524bf04b7ab"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:18:38 crc kubenswrapper[4873]: I1201 09:18:38.064102 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99a70cc5-d0e6-4066-9fdb-7524bf04b7ab-inventory" (OuterVolumeSpecName: "inventory") pod "99a70cc5-d0e6-4066-9fdb-7524bf04b7ab" (UID: "99a70cc5-d0e6-4066-9fdb-7524bf04b7ab"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:18:38 crc kubenswrapper[4873]: I1201 09:18:38.127790 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zb97c\" (UniqueName: \"kubernetes.io/projected/99a70cc5-d0e6-4066-9fdb-7524bf04b7ab-kube-api-access-zb97c\") on node \"crc\" DevicePath \"\"" Dec 01 09:18:38 crc kubenswrapper[4873]: I1201 09:18:38.127837 4873 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/99a70cc5-d0e6-4066-9fdb-7524bf04b7ab-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 09:18:38 crc kubenswrapper[4873]: I1201 09:18:38.127846 4873 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/99a70cc5-d0e6-4066-9fdb-7524bf04b7ab-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 09:18:38 crc kubenswrapper[4873]: I1201 09:18:38.127855 4873 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/99a70cc5-d0e6-4066-9fdb-7524bf04b7ab-ceph\") on node \"crc\" DevicePath \"\"" Dec 01 09:18:38 crc kubenswrapper[4873]: I1201 09:18:38.378924 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-lcck8" event={"ID":"99a70cc5-d0e6-4066-9fdb-7524bf04b7ab","Type":"ContainerDied","Data":"3223ddd4dc8bbb4e79a040ecafc989178550c5c275d093617ede481e4c32fd82"} Dec 01 09:18:38 crc kubenswrapper[4873]: I1201 09:18:38.378976 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3223ddd4dc8bbb4e79a040ecafc989178550c5c275d093617ede481e4c32fd82" Dec 01 09:18:38 crc kubenswrapper[4873]: I1201 09:18:38.378985 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-lcck8" Dec 01 09:18:38 crc kubenswrapper[4873]: I1201 09:18:38.463410 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qt48t"] Dec 01 09:18:38 crc kubenswrapper[4873]: E1201 09:18:38.463876 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99a70cc5-d0e6-4066-9fdb-7524bf04b7ab" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 01 09:18:38 crc kubenswrapper[4873]: I1201 09:18:38.463899 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="99a70cc5-d0e6-4066-9fdb-7524bf04b7ab" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 01 09:18:38 crc kubenswrapper[4873]: I1201 09:18:38.464125 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="99a70cc5-d0e6-4066-9fdb-7524bf04b7ab" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 01 09:18:38 crc kubenswrapper[4873]: I1201 09:18:38.464828 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qt48t" Dec 01 09:18:38 crc kubenswrapper[4873]: I1201 09:18:38.468654 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 01 09:18:38 crc kubenswrapper[4873]: I1201 09:18:38.468963 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 09:18:38 crc kubenswrapper[4873]: I1201 09:18:38.469208 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 09:18:38 crc kubenswrapper[4873]: I1201 09:18:38.473042 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 09:18:38 crc kubenswrapper[4873]: I1201 09:18:38.473091 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9ghqw" Dec 01 09:18:38 crc kubenswrapper[4873]: I1201 09:18:38.480197 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qt48t"] Dec 01 09:18:38 crc kubenswrapper[4873]: I1201 09:18:38.535375 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2xkl\" (UniqueName: \"kubernetes.io/projected/933df0ea-eae8-49a3-affc-2f421e3ae777-kube-api-access-j2xkl\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qt48t\" (UID: \"933df0ea-eae8-49a3-affc-2f421e3ae777\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qt48t" Dec 01 09:18:38 crc kubenswrapper[4873]: I1201 09:18:38.535664 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/933df0ea-eae8-49a3-affc-2f421e3ae777-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qt48t\" (UID: \"933df0ea-eae8-49a3-affc-2f421e3ae777\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qt48t" Dec 01 09:18:38 crc kubenswrapper[4873]: I1201 09:18:38.535742 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/933df0ea-eae8-49a3-affc-2f421e3ae777-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qt48t\" (UID: \"933df0ea-eae8-49a3-affc-2f421e3ae777\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qt48t" Dec 01 09:18:38 crc kubenswrapper[4873]: I1201 09:18:38.535993 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/933df0ea-eae8-49a3-affc-2f421e3ae777-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qt48t\" (UID: \"933df0ea-eae8-49a3-affc-2f421e3ae777\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qt48t" Dec 01 09:18:38 crc kubenswrapper[4873]: I1201 09:18:38.637858 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/933df0ea-eae8-49a3-affc-2f421e3ae777-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qt48t\" (UID: \"933df0ea-eae8-49a3-affc-2f421e3ae777\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qt48t" Dec 01 09:18:38 crc kubenswrapper[4873]: I1201 09:18:38.637918 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/933df0ea-eae8-49a3-affc-2f421e3ae777-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qt48t\" (UID: \"933df0ea-eae8-49a3-affc-2f421e3ae777\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qt48t" Dec 01 09:18:38 crc kubenswrapper[4873]: I1201 09:18:38.637996 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/933df0ea-eae8-49a3-affc-2f421e3ae777-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qt48t\" (UID: \"933df0ea-eae8-49a3-affc-2f421e3ae777\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qt48t" Dec 01 09:18:38 crc kubenswrapper[4873]: I1201 09:18:38.638070 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2xkl\" (UniqueName: \"kubernetes.io/projected/933df0ea-eae8-49a3-affc-2f421e3ae777-kube-api-access-j2xkl\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qt48t\" (UID: \"933df0ea-eae8-49a3-affc-2f421e3ae777\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qt48t" Dec 01 09:18:38 crc kubenswrapper[4873]: I1201 09:18:38.643813 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/933df0ea-eae8-49a3-affc-2f421e3ae777-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qt48t\" (UID: \"933df0ea-eae8-49a3-affc-2f421e3ae777\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qt48t" Dec 01 09:18:38 crc kubenswrapper[4873]: I1201 09:18:38.644434 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/933df0ea-eae8-49a3-affc-2f421e3ae777-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qt48t\" (UID: \"933df0ea-eae8-49a3-affc-2f421e3ae777\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qt48t" Dec 01 09:18:38 crc kubenswrapper[4873]: I1201 09:18:38.644599 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/933df0ea-eae8-49a3-affc-2f421e3ae777-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qt48t\" (UID: \"933df0ea-eae8-49a3-affc-2f421e3ae777\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qt48t" Dec 01 09:18:38 crc kubenswrapper[4873]: I1201 09:18:38.659954 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2xkl\" (UniqueName: \"kubernetes.io/projected/933df0ea-eae8-49a3-affc-2f421e3ae777-kube-api-access-j2xkl\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qt48t\" (UID: \"933df0ea-eae8-49a3-affc-2f421e3ae777\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qt48t" Dec 01 09:18:38 crc kubenswrapper[4873]: I1201 09:18:38.795524 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qt48t" Dec 01 09:18:39 crc kubenswrapper[4873]: I1201 09:18:39.366291 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qt48t"] Dec 01 09:18:39 crc kubenswrapper[4873]: I1201 09:18:39.388816 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qt48t" event={"ID":"933df0ea-eae8-49a3-affc-2f421e3ae777","Type":"ContainerStarted","Data":"3e01744ac8e69af0b49e9c9e0417e1dc103f33fcc22ff7c82b0d06d38934901c"} Dec 01 09:18:40 crc kubenswrapper[4873]: I1201 09:18:40.414515 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qt48t" event={"ID":"933df0ea-eae8-49a3-affc-2f421e3ae777","Type":"ContainerStarted","Data":"e31dea2961dae62bde49564b4c217fe832391facd60c69f605ac9100fc874161"} Dec 01 09:18:40 crc kubenswrapper[4873]: I1201 09:18:40.462330 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qt48t" podStartSLOduration=1.984204472 podStartE2EDuration="2.462298718s" podCreationTimestamp="2025-12-01 09:18:38 +0000 UTC" firstStartedPulling="2025-12-01 09:18:39.378638201 +0000 UTC m=+2295.280746740" lastFinishedPulling="2025-12-01 09:18:39.856732427 +0000 UTC m=+2295.758840986" observedRunningTime="2025-12-01 09:18:40.44818685 +0000 UTC m=+2296.350295399" watchObservedRunningTime="2025-12-01 09:18:40.462298718 +0000 UTC m=+2296.364407267" Dec 01 09:18:42 crc kubenswrapper[4873]: I1201 09:18:42.431115 4873 scope.go:117] "RemoveContainer" containerID="1a1db8f8327e8cf6dbb94b57ea898b250339bb198ae1ba595cb1d79805fa0db1" Dec 01 09:18:42 crc kubenswrapper[4873]: E1201 09:18:42.432165 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:18:45 crc kubenswrapper[4873]: I1201 09:18:45.462340 4873 generic.go:334] "Generic (PLEG): container finished" podID="933df0ea-eae8-49a3-affc-2f421e3ae777" containerID="e31dea2961dae62bde49564b4c217fe832391facd60c69f605ac9100fc874161" exitCode=0 Dec 01 09:18:45 crc kubenswrapper[4873]: I1201 09:18:45.462429 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qt48t" event={"ID":"933df0ea-eae8-49a3-affc-2f421e3ae777","Type":"ContainerDied","Data":"e31dea2961dae62bde49564b4c217fe832391facd60c69f605ac9100fc874161"} Dec 01 09:18:46 crc kubenswrapper[4873]: I1201 09:18:46.914036 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qt48t" Dec 01 09:18:46 crc kubenswrapper[4873]: I1201 09:18:46.922496 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2xkl\" (UniqueName: \"kubernetes.io/projected/933df0ea-eae8-49a3-affc-2f421e3ae777-kube-api-access-j2xkl\") pod \"933df0ea-eae8-49a3-affc-2f421e3ae777\" (UID: \"933df0ea-eae8-49a3-affc-2f421e3ae777\") " Dec 01 09:18:46 crc kubenswrapper[4873]: I1201 09:18:46.922684 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/933df0ea-eae8-49a3-affc-2f421e3ae777-inventory\") pod \"933df0ea-eae8-49a3-affc-2f421e3ae777\" (UID: \"933df0ea-eae8-49a3-affc-2f421e3ae777\") " Dec 01 09:18:46 crc kubenswrapper[4873]: I1201 09:18:46.922729 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/933df0ea-eae8-49a3-affc-2f421e3ae777-ceph\") pod \"933df0ea-eae8-49a3-affc-2f421e3ae777\" (UID: \"933df0ea-eae8-49a3-affc-2f421e3ae777\") " Dec 01 09:18:46 crc kubenswrapper[4873]: I1201 09:18:46.922823 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/933df0ea-eae8-49a3-affc-2f421e3ae777-ssh-key\") pod \"933df0ea-eae8-49a3-affc-2f421e3ae777\" (UID: \"933df0ea-eae8-49a3-affc-2f421e3ae777\") " Dec 01 09:18:46 crc kubenswrapper[4873]: I1201 09:18:46.929294 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/933df0ea-eae8-49a3-affc-2f421e3ae777-ceph" (OuterVolumeSpecName: "ceph") pod "933df0ea-eae8-49a3-affc-2f421e3ae777" (UID: "933df0ea-eae8-49a3-affc-2f421e3ae777"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:18:46 crc kubenswrapper[4873]: I1201 09:18:46.930206 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/933df0ea-eae8-49a3-affc-2f421e3ae777-kube-api-access-j2xkl" (OuterVolumeSpecName: "kube-api-access-j2xkl") pod "933df0ea-eae8-49a3-affc-2f421e3ae777" (UID: "933df0ea-eae8-49a3-affc-2f421e3ae777"). InnerVolumeSpecName "kube-api-access-j2xkl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:18:46 crc kubenswrapper[4873]: I1201 09:18:46.959125 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/933df0ea-eae8-49a3-affc-2f421e3ae777-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "933df0ea-eae8-49a3-affc-2f421e3ae777" (UID: "933df0ea-eae8-49a3-affc-2f421e3ae777"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:18:46 crc kubenswrapper[4873]: I1201 09:18:46.963527 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/933df0ea-eae8-49a3-affc-2f421e3ae777-inventory" (OuterVolumeSpecName: "inventory") pod "933df0ea-eae8-49a3-affc-2f421e3ae777" (UID: "933df0ea-eae8-49a3-affc-2f421e3ae777"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:18:47 crc kubenswrapper[4873]: I1201 09:18:47.025431 4873 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/933df0ea-eae8-49a3-affc-2f421e3ae777-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 09:18:47 crc kubenswrapper[4873]: I1201 09:18:47.025935 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2xkl\" (UniqueName: \"kubernetes.io/projected/933df0ea-eae8-49a3-affc-2f421e3ae777-kube-api-access-j2xkl\") on node \"crc\" DevicePath \"\"" Dec 01 09:18:47 crc kubenswrapper[4873]: I1201 09:18:47.025957 4873 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/933df0ea-eae8-49a3-affc-2f421e3ae777-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 09:18:47 crc kubenswrapper[4873]: I1201 09:18:47.025968 4873 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/933df0ea-eae8-49a3-affc-2f421e3ae777-ceph\") on node \"crc\" DevicePath \"\"" Dec 01 09:18:47 crc kubenswrapper[4873]: I1201 09:18:47.482319 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qt48t" event={"ID":"933df0ea-eae8-49a3-affc-2f421e3ae777","Type":"ContainerDied","Data":"3e01744ac8e69af0b49e9c9e0417e1dc103f33fcc22ff7c82b0d06d38934901c"} Dec 01 09:18:47 crc kubenswrapper[4873]: I1201 09:18:47.482371 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3e01744ac8e69af0b49e9c9e0417e1dc103f33fcc22ff7c82b0d06d38934901c" Dec 01 09:18:47 crc kubenswrapper[4873]: I1201 09:18:47.482439 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qt48t" Dec 01 09:18:47 crc kubenswrapper[4873]: I1201 09:18:47.686892 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-99878"] Dec 01 09:18:47 crc kubenswrapper[4873]: E1201 09:18:47.687462 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="933df0ea-eae8-49a3-affc-2f421e3ae777" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 01 09:18:47 crc kubenswrapper[4873]: I1201 09:18:47.687494 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="933df0ea-eae8-49a3-affc-2f421e3ae777" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 01 09:18:47 crc kubenswrapper[4873]: I1201 09:18:47.687757 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="933df0ea-eae8-49a3-affc-2f421e3ae777" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 01 09:18:47 crc kubenswrapper[4873]: I1201 09:18:47.689351 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-99878" Dec 01 09:18:47 crc kubenswrapper[4873]: I1201 09:18:47.692266 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 09:18:47 crc kubenswrapper[4873]: I1201 09:18:47.692313 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9ghqw" Dec 01 09:18:47 crc kubenswrapper[4873]: I1201 09:18:47.692835 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 01 09:18:47 crc kubenswrapper[4873]: I1201 09:18:47.694920 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 09:18:47 crc kubenswrapper[4873]: I1201 09:18:47.695586 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 09:18:47 crc kubenswrapper[4873]: I1201 09:18:47.700451 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-99878"] Dec 01 09:18:47 crc kubenswrapper[4873]: I1201 09:18:47.702871 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkhpv\" (UniqueName: \"kubernetes.io/projected/83908788-7c1f-410a-be4d-79510d6703b5-kube-api-access-xkhpv\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-99878\" (UID: \"83908788-7c1f-410a-be4d-79510d6703b5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-99878" Dec 01 09:18:47 crc kubenswrapper[4873]: I1201 09:18:47.702932 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/83908788-7c1f-410a-be4d-79510d6703b5-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-99878\" (UID: \"83908788-7c1f-410a-be4d-79510d6703b5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-99878" Dec 01 09:18:47 crc kubenswrapper[4873]: I1201 09:18:47.702960 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/83908788-7c1f-410a-be4d-79510d6703b5-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-99878\" (UID: \"83908788-7c1f-410a-be4d-79510d6703b5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-99878" Dec 01 09:18:47 crc kubenswrapper[4873]: I1201 09:18:47.703263 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/83908788-7c1f-410a-be4d-79510d6703b5-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-99878\" (UID: \"83908788-7c1f-410a-be4d-79510d6703b5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-99878" Dec 01 09:18:47 crc kubenswrapper[4873]: I1201 09:18:47.805234 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/83908788-7c1f-410a-be4d-79510d6703b5-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-99878\" (UID: \"83908788-7c1f-410a-be4d-79510d6703b5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-99878" Dec 01 09:18:47 crc kubenswrapper[4873]: I1201 09:18:47.805341 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkhpv\" (UniqueName: \"kubernetes.io/projected/83908788-7c1f-410a-be4d-79510d6703b5-kube-api-access-xkhpv\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-99878\" (UID: \"83908788-7c1f-410a-be4d-79510d6703b5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-99878" Dec 01 09:18:47 crc kubenswrapper[4873]: I1201 09:18:47.805380 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/83908788-7c1f-410a-be4d-79510d6703b5-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-99878\" (UID: \"83908788-7c1f-410a-be4d-79510d6703b5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-99878" Dec 01 09:18:47 crc kubenswrapper[4873]: I1201 09:18:47.805414 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/83908788-7c1f-410a-be4d-79510d6703b5-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-99878\" (UID: \"83908788-7c1f-410a-be4d-79510d6703b5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-99878" Dec 01 09:18:47 crc kubenswrapper[4873]: I1201 09:18:47.812741 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/83908788-7c1f-410a-be4d-79510d6703b5-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-99878\" (UID: \"83908788-7c1f-410a-be4d-79510d6703b5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-99878" Dec 01 09:18:47 crc kubenswrapper[4873]: I1201 09:18:47.822728 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/83908788-7c1f-410a-be4d-79510d6703b5-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-99878\" (UID: \"83908788-7c1f-410a-be4d-79510d6703b5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-99878" Dec 01 09:18:47 crc kubenswrapper[4873]: I1201 09:18:47.823356 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/83908788-7c1f-410a-be4d-79510d6703b5-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-99878\" (UID: \"83908788-7c1f-410a-be4d-79510d6703b5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-99878" Dec 01 09:18:47 crc kubenswrapper[4873]: I1201 09:18:47.825468 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkhpv\" (UniqueName: \"kubernetes.io/projected/83908788-7c1f-410a-be4d-79510d6703b5-kube-api-access-xkhpv\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-99878\" (UID: \"83908788-7c1f-410a-be4d-79510d6703b5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-99878" Dec 01 09:18:48 crc kubenswrapper[4873]: I1201 09:18:48.010931 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-99878" Dec 01 09:18:48 crc kubenswrapper[4873]: I1201 09:18:48.585569 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-99878"] Dec 01 09:18:49 crc kubenswrapper[4873]: I1201 09:18:49.501714 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-99878" event={"ID":"83908788-7c1f-410a-be4d-79510d6703b5","Type":"ContainerStarted","Data":"2c2ed08ac001aea2970b0e4c0d591c1d82204ce4af13413234273709d46a169c"} Dec 01 09:18:50 crc kubenswrapper[4873]: I1201 09:18:50.514432 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-99878" event={"ID":"83908788-7c1f-410a-be4d-79510d6703b5","Type":"ContainerStarted","Data":"df2d453fcb568766c673d651313645471da3ecc782dd06b9f2d07adc01e8f2f9"} Dec 01 09:18:50 crc kubenswrapper[4873]: I1201 09:18:50.538489 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-99878" podStartSLOduration=2.732252469 podStartE2EDuration="3.538437641s" podCreationTimestamp="2025-12-01 09:18:47 +0000 UTC" firstStartedPulling="2025-12-01 09:18:48.595114422 +0000 UTC m=+2304.497222951" lastFinishedPulling="2025-12-01 09:18:49.401299584 +0000 UTC m=+2305.303408123" observedRunningTime="2025-12-01 09:18:50.535860398 +0000 UTC m=+2306.437968937" watchObservedRunningTime="2025-12-01 09:18:50.538437641 +0000 UTC m=+2306.440546180" Dec 01 09:18:53 crc kubenswrapper[4873]: I1201 09:18:53.431108 4873 scope.go:117] "RemoveContainer" containerID="1a1db8f8327e8cf6dbb94b57ea898b250339bb198ae1ba595cb1d79805fa0db1" Dec 01 09:18:53 crc kubenswrapper[4873]: E1201 09:18:53.431893 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:19:05 crc kubenswrapper[4873]: I1201 09:19:05.430988 4873 scope.go:117] "RemoveContainer" containerID="1a1db8f8327e8cf6dbb94b57ea898b250339bb198ae1ba595cb1d79805fa0db1" Dec 01 09:19:05 crc kubenswrapper[4873]: E1201 09:19:05.432413 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:19:17 crc kubenswrapper[4873]: I1201 09:19:17.430115 4873 scope.go:117] "RemoveContainer" containerID="1a1db8f8327e8cf6dbb94b57ea898b250339bb198ae1ba595cb1d79805fa0db1" Dec 01 09:19:17 crc kubenswrapper[4873]: E1201 09:19:17.431433 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:19:26 crc kubenswrapper[4873]: I1201 09:19:26.845277 4873 generic.go:334] "Generic (PLEG): container finished" podID="83908788-7c1f-410a-be4d-79510d6703b5" containerID="df2d453fcb568766c673d651313645471da3ecc782dd06b9f2d07adc01e8f2f9" exitCode=0 Dec 01 09:19:26 crc kubenswrapper[4873]: I1201 09:19:26.845367 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-99878" event={"ID":"83908788-7c1f-410a-be4d-79510d6703b5","Type":"ContainerDied","Data":"df2d453fcb568766c673d651313645471da3ecc782dd06b9f2d07adc01e8f2f9"} Dec 01 09:19:28 crc kubenswrapper[4873]: I1201 09:19:28.286298 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-99878" Dec 01 09:19:28 crc kubenswrapper[4873]: I1201 09:19:28.387133 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/83908788-7c1f-410a-be4d-79510d6703b5-ceph\") pod \"83908788-7c1f-410a-be4d-79510d6703b5\" (UID: \"83908788-7c1f-410a-be4d-79510d6703b5\") " Dec 01 09:19:28 crc kubenswrapper[4873]: I1201 09:19:28.387303 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/83908788-7c1f-410a-be4d-79510d6703b5-ssh-key\") pod \"83908788-7c1f-410a-be4d-79510d6703b5\" (UID: \"83908788-7c1f-410a-be4d-79510d6703b5\") " Dec 01 09:19:28 crc kubenswrapper[4873]: I1201 09:19:28.387474 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xkhpv\" (UniqueName: \"kubernetes.io/projected/83908788-7c1f-410a-be4d-79510d6703b5-kube-api-access-xkhpv\") pod \"83908788-7c1f-410a-be4d-79510d6703b5\" (UID: \"83908788-7c1f-410a-be4d-79510d6703b5\") " Dec 01 09:19:28 crc kubenswrapper[4873]: I1201 09:19:28.387593 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/83908788-7c1f-410a-be4d-79510d6703b5-inventory\") pod \"83908788-7c1f-410a-be4d-79510d6703b5\" (UID: \"83908788-7c1f-410a-be4d-79510d6703b5\") " Dec 01 09:19:28 crc kubenswrapper[4873]: I1201 09:19:28.395841 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83908788-7c1f-410a-be4d-79510d6703b5-ceph" (OuterVolumeSpecName: "ceph") pod "83908788-7c1f-410a-be4d-79510d6703b5" (UID: "83908788-7c1f-410a-be4d-79510d6703b5"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:19:28 crc kubenswrapper[4873]: I1201 09:19:28.396271 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83908788-7c1f-410a-be4d-79510d6703b5-kube-api-access-xkhpv" (OuterVolumeSpecName: "kube-api-access-xkhpv") pod "83908788-7c1f-410a-be4d-79510d6703b5" (UID: "83908788-7c1f-410a-be4d-79510d6703b5"). InnerVolumeSpecName "kube-api-access-xkhpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:19:28 crc kubenswrapper[4873]: I1201 09:19:28.422683 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83908788-7c1f-410a-be4d-79510d6703b5-inventory" (OuterVolumeSpecName: "inventory") pod "83908788-7c1f-410a-be4d-79510d6703b5" (UID: "83908788-7c1f-410a-be4d-79510d6703b5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:19:28 crc kubenswrapper[4873]: I1201 09:19:28.433648 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83908788-7c1f-410a-be4d-79510d6703b5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "83908788-7c1f-410a-be4d-79510d6703b5" (UID: "83908788-7c1f-410a-be4d-79510d6703b5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:19:28 crc kubenswrapper[4873]: I1201 09:19:28.489911 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xkhpv\" (UniqueName: \"kubernetes.io/projected/83908788-7c1f-410a-be4d-79510d6703b5-kube-api-access-xkhpv\") on node \"crc\" DevicePath \"\"" Dec 01 09:19:28 crc kubenswrapper[4873]: I1201 09:19:28.490002 4873 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/83908788-7c1f-410a-be4d-79510d6703b5-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 09:19:28 crc kubenswrapper[4873]: I1201 09:19:28.490033 4873 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/83908788-7c1f-410a-be4d-79510d6703b5-ceph\") on node \"crc\" DevicePath \"\"" Dec 01 09:19:28 crc kubenswrapper[4873]: I1201 09:19:28.490044 4873 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/83908788-7c1f-410a-be4d-79510d6703b5-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 09:19:28 crc kubenswrapper[4873]: I1201 09:19:28.866987 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-99878" event={"ID":"83908788-7c1f-410a-be4d-79510d6703b5","Type":"ContainerDied","Data":"2c2ed08ac001aea2970b0e4c0d591c1d82204ce4af13413234273709d46a169c"} Dec 01 09:19:28 crc kubenswrapper[4873]: I1201 09:19:28.867066 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2c2ed08ac001aea2970b0e4c0d591c1d82204ce4af13413234273709d46a169c" Dec 01 09:19:28 crc kubenswrapper[4873]: I1201 09:19:28.867039 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-99878" Dec 01 09:19:28 crc kubenswrapper[4873]: I1201 09:19:28.965396 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-75xtr"] Dec 01 09:19:28 crc kubenswrapper[4873]: E1201 09:19:28.966036 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83908788-7c1f-410a-be4d-79510d6703b5" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 01 09:19:28 crc kubenswrapper[4873]: I1201 09:19:28.966061 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="83908788-7c1f-410a-be4d-79510d6703b5" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 01 09:19:28 crc kubenswrapper[4873]: I1201 09:19:28.966319 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="83908788-7c1f-410a-be4d-79510d6703b5" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 01 09:19:28 crc kubenswrapper[4873]: I1201 09:19:28.967385 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-75xtr" Dec 01 09:19:28 crc kubenswrapper[4873]: I1201 09:19:28.970893 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 09:19:28 crc kubenswrapper[4873]: I1201 09:19:28.971163 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 01 09:19:28 crc kubenswrapper[4873]: I1201 09:19:28.971488 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 09:19:28 crc kubenswrapper[4873]: I1201 09:19:28.971716 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 09:19:28 crc kubenswrapper[4873]: I1201 09:19:28.974782 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9ghqw" Dec 01 09:19:28 crc kubenswrapper[4873]: I1201 09:19:28.976885 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-75xtr"] Dec 01 09:19:29 crc kubenswrapper[4873]: I1201 09:19:29.106263 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b355b9ce-a737-407d-8c9e-5d9db024bb10-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-75xtr\" (UID: \"b355b9ce-a737-407d-8c9e-5d9db024bb10\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-75xtr" Dec 01 09:19:29 crc kubenswrapper[4873]: I1201 09:19:29.106353 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b355b9ce-a737-407d-8c9e-5d9db024bb10-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-75xtr\" (UID: \"b355b9ce-a737-407d-8c9e-5d9db024bb10\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-75xtr" Dec 01 09:19:29 crc kubenswrapper[4873]: I1201 09:19:29.106381 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-477bs\" (UniqueName: \"kubernetes.io/projected/b355b9ce-a737-407d-8c9e-5d9db024bb10-kube-api-access-477bs\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-75xtr\" (UID: \"b355b9ce-a737-407d-8c9e-5d9db024bb10\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-75xtr" Dec 01 09:19:29 crc kubenswrapper[4873]: I1201 09:19:29.106409 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b355b9ce-a737-407d-8c9e-5d9db024bb10-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-75xtr\" (UID: \"b355b9ce-a737-407d-8c9e-5d9db024bb10\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-75xtr" Dec 01 09:19:29 crc kubenswrapper[4873]: I1201 09:19:29.208859 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b355b9ce-a737-407d-8c9e-5d9db024bb10-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-75xtr\" (UID: \"b355b9ce-a737-407d-8c9e-5d9db024bb10\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-75xtr" Dec 01 09:19:29 crc kubenswrapper[4873]: I1201 09:19:29.208944 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b355b9ce-a737-407d-8c9e-5d9db024bb10-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-75xtr\" (UID: \"b355b9ce-a737-407d-8c9e-5d9db024bb10\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-75xtr" Dec 01 09:19:29 crc kubenswrapper[4873]: I1201 09:19:29.208984 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-477bs\" (UniqueName: \"kubernetes.io/projected/b355b9ce-a737-407d-8c9e-5d9db024bb10-kube-api-access-477bs\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-75xtr\" (UID: \"b355b9ce-a737-407d-8c9e-5d9db024bb10\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-75xtr" Dec 01 09:19:29 crc kubenswrapper[4873]: I1201 09:19:29.209040 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b355b9ce-a737-407d-8c9e-5d9db024bb10-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-75xtr\" (UID: \"b355b9ce-a737-407d-8c9e-5d9db024bb10\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-75xtr" Dec 01 09:19:29 crc kubenswrapper[4873]: I1201 09:19:29.214612 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b355b9ce-a737-407d-8c9e-5d9db024bb10-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-75xtr\" (UID: \"b355b9ce-a737-407d-8c9e-5d9db024bb10\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-75xtr" Dec 01 09:19:29 crc kubenswrapper[4873]: I1201 09:19:29.215088 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b355b9ce-a737-407d-8c9e-5d9db024bb10-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-75xtr\" (UID: \"b355b9ce-a737-407d-8c9e-5d9db024bb10\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-75xtr" Dec 01 09:19:29 crc kubenswrapper[4873]: I1201 09:19:29.215276 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b355b9ce-a737-407d-8c9e-5d9db024bb10-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-75xtr\" (UID: \"b355b9ce-a737-407d-8c9e-5d9db024bb10\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-75xtr" Dec 01 09:19:29 crc kubenswrapper[4873]: I1201 09:19:29.232632 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-477bs\" (UniqueName: \"kubernetes.io/projected/b355b9ce-a737-407d-8c9e-5d9db024bb10-kube-api-access-477bs\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-75xtr\" (UID: \"b355b9ce-a737-407d-8c9e-5d9db024bb10\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-75xtr" Dec 01 09:19:29 crc kubenswrapper[4873]: I1201 09:19:29.293908 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-75xtr" Dec 01 09:19:29 crc kubenswrapper[4873]: I1201 09:19:29.656356 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-75xtr"] Dec 01 09:19:29 crc kubenswrapper[4873]: I1201 09:19:29.878324 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-75xtr" event={"ID":"b355b9ce-a737-407d-8c9e-5d9db024bb10","Type":"ContainerStarted","Data":"a3f513629efaf3eef44615ad99d022d6b36b3393a75e6d83f269454e1181e1fe"} Dec 01 09:19:30 crc kubenswrapper[4873]: I1201 09:19:30.431929 4873 scope.go:117] "RemoveContainer" containerID="1a1db8f8327e8cf6dbb94b57ea898b250339bb198ae1ba595cb1d79805fa0db1" Dec 01 09:19:30 crc kubenswrapper[4873]: E1201 09:19:30.432723 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:19:30 crc kubenswrapper[4873]: I1201 09:19:30.890383 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-75xtr" event={"ID":"b355b9ce-a737-407d-8c9e-5d9db024bb10","Type":"ContainerStarted","Data":"b591f4911995027f4c6992386f45af85d7453a1ef8ad3871ec6b26bb1b430f4b"} Dec 01 09:19:30 crc kubenswrapper[4873]: I1201 09:19:30.921548 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-75xtr" podStartSLOduration=2.296885959 podStartE2EDuration="2.921515768s" podCreationTimestamp="2025-12-01 09:19:28 +0000 UTC" firstStartedPulling="2025-12-01 09:19:29.665245327 +0000 UTC m=+2345.567353876" lastFinishedPulling="2025-12-01 09:19:30.289875146 +0000 UTC m=+2346.191983685" observedRunningTime="2025-12-01 09:19:30.910211219 +0000 UTC m=+2346.812319778" watchObservedRunningTime="2025-12-01 09:19:30.921515768 +0000 UTC m=+2346.823624307" Dec 01 09:19:34 crc kubenswrapper[4873]: I1201 09:19:34.930113 4873 generic.go:334] "Generic (PLEG): container finished" podID="b355b9ce-a737-407d-8c9e-5d9db024bb10" containerID="b591f4911995027f4c6992386f45af85d7453a1ef8ad3871ec6b26bb1b430f4b" exitCode=0 Dec 01 09:19:34 crc kubenswrapper[4873]: I1201 09:19:34.930213 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-75xtr" event={"ID":"b355b9ce-a737-407d-8c9e-5d9db024bb10","Type":"ContainerDied","Data":"b591f4911995027f4c6992386f45af85d7453a1ef8ad3871ec6b26bb1b430f4b"} Dec 01 09:19:36 crc kubenswrapper[4873]: I1201 09:19:36.413379 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-75xtr" Dec 01 09:19:36 crc kubenswrapper[4873]: I1201 09:19:36.526560 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b355b9ce-a737-407d-8c9e-5d9db024bb10-inventory\") pod \"b355b9ce-a737-407d-8c9e-5d9db024bb10\" (UID: \"b355b9ce-a737-407d-8c9e-5d9db024bb10\") " Dec 01 09:19:36 crc kubenswrapper[4873]: I1201 09:19:36.526643 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-477bs\" (UniqueName: \"kubernetes.io/projected/b355b9ce-a737-407d-8c9e-5d9db024bb10-kube-api-access-477bs\") pod \"b355b9ce-a737-407d-8c9e-5d9db024bb10\" (UID: \"b355b9ce-a737-407d-8c9e-5d9db024bb10\") " Dec 01 09:19:36 crc kubenswrapper[4873]: I1201 09:19:36.526788 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b355b9ce-a737-407d-8c9e-5d9db024bb10-ceph\") pod \"b355b9ce-a737-407d-8c9e-5d9db024bb10\" (UID: \"b355b9ce-a737-407d-8c9e-5d9db024bb10\") " Dec 01 09:19:36 crc kubenswrapper[4873]: I1201 09:19:36.526909 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b355b9ce-a737-407d-8c9e-5d9db024bb10-ssh-key\") pod \"b355b9ce-a737-407d-8c9e-5d9db024bb10\" (UID: \"b355b9ce-a737-407d-8c9e-5d9db024bb10\") " Dec 01 09:19:36 crc kubenswrapper[4873]: I1201 09:19:36.533004 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b355b9ce-a737-407d-8c9e-5d9db024bb10-kube-api-access-477bs" (OuterVolumeSpecName: "kube-api-access-477bs") pod "b355b9ce-a737-407d-8c9e-5d9db024bb10" (UID: "b355b9ce-a737-407d-8c9e-5d9db024bb10"). InnerVolumeSpecName "kube-api-access-477bs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:19:36 crc kubenswrapper[4873]: I1201 09:19:36.533637 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b355b9ce-a737-407d-8c9e-5d9db024bb10-ceph" (OuterVolumeSpecName: "ceph") pod "b355b9ce-a737-407d-8c9e-5d9db024bb10" (UID: "b355b9ce-a737-407d-8c9e-5d9db024bb10"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:19:36 crc kubenswrapper[4873]: I1201 09:19:36.559505 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b355b9ce-a737-407d-8c9e-5d9db024bb10-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b355b9ce-a737-407d-8c9e-5d9db024bb10" (UID: "b355b9ce-a737-407d-8c9e-5d9db024bb10"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:19:36 crc kubenswrapper[4873]: I1201 09:19:36.559593 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b355b9ce-a737-407d-8c9e-5d9db024bb10-inventory" (OuterVolumeSpecName: "inventory") pod "b355b9ce-a737-407d-8c9e-5d9db024bb10" (UID: "b355b9ce-a737-407d-8c9e-5d9db024bb10"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:19:36 crc kubenswrapper[4873]: I1201 09:19:36.631542 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-477bs\" (UniqueName: \"kubernetes.io/projected/b355b9ce-a737-407d-8c9e-5d9db024bb10-kube-api-access-477bs\") on node \"crc\" DevicePath \"\"" Dec 01 09:19:36 crc kubenswrapper[4873]: I1201 09:19:36.631599 4873 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b355b9ce-a737-407d-8c9e-5d9db024bb10-ceph\") on node \"crc\" DevicePath \"\"" Dec 01 09:19:36 crc kubenswrapper[4873]: I1201 09:19:36.631612 4873 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b355b9ce-a737-407d-8c9e-5d9db024bb10-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 09:19:36 crc kubenswrapper[4873]: I1201 09:19:36.631623 4873 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b355b9ce-a737-407d-8c9e-5d9db024bb10-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 09:19:36 crc kubenswrapper[4873]: I1201 09:19:36.950211 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-75xtr" event={"ID":"b355b9ce-a737-407d-8c9e-5d9db024bb10","Type":"ContainerDied","Data":"a3f513629efaf3eef44615ad99d022d6b36b3393a75e6d83f269454e1181e1fe"} Dec 01 09:19:36 crc kubenswrapper[4873]: I1201 09:19:36.950598 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a3f513629efaf3eef44615ad99d022d6b36b3393a75e6d83f269454e1181e1fe" Dec 01 09:19:36 crc kubenswrapper[4873]: I1201 09:19:36.950296 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-75xtr" Dec 01 09:19:37 crc kubenswrapper[4873]: I1201 09:19:37.040448 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kntht"] Dec 01 09:19:37 crc kubenswrapper[4873]: E1201 09:19:37.041148 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b355b9ce-a737-407d-8c9e-5d9db024bb10" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 01 09:19:37 crc kubenswrapper[4873]: I1201 09:19:37.041272 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="b355b9ce-a737-407d-8c9e-5d9db024bb10" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 01 09:19:37 crc kubenswrapper[4873]: I1201 09:19:37.041562 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="b355b9ce-a737-407d-8c9e-5d9db024bb10" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 01 09:19:37 crc kubenswrapper[4873]: I1201 09:19:37.042342 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kntht" Dec 01 09:19:37 crc kubenswrapper[4873]: I1201 09:19:37.046858 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 09:19:37 crc kubenswrapper[4873]: I1201 09:19:37.047064 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 09:19:37 crc kubenswrapper[4873]: I1201 09:19:37.046961 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 09:19:37 crc kubenswrapper[4873]: I1201 09:19:37.047008 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 01 09:19:37 crc kubenswrapper[4873]: I1201 09:19:37.047400 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9ghqw" Dec 01 09:19:37 crc kubenswrapper[4873]: I1201 09:19:37.071945 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kntht"] Dec 01 09:19:37 crc kubenswrapper[4873]: I1201 09:19:37.142869 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqrnt\" (UniqueName: \"kubernetes.io/projected/cf256dda-e402-4a9f-bff1-fe2990f7ce72-kube-api-access-zqrnt\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kntht\" (UID: \"cf256dda-e402-4a9f-bff1-fe2990f7ce72\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kntht" Dec 01 09:19:37 crc kubenswrapper[4873]: I1201 09:19:37.142927 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cf256dda-e402-4a9f-bff1-fe2990f7ce72-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kntht\" (UID: \"cf256dda-e402-4a9f-bff1-fe2990f7ce72\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kntht" Dec 01 09:19:37 crc kubenswrapper[4873]: I1201 09:19:37.142963 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cf256dda-e402-4a9f-bff1-fe2990f7ce72-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kntht\" (UID: \"cf256dda-e402-4a9f-bff1-fe2990f7ce72\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kntht" Dec 01 09:19:37 crc kubenswrapper[4873]: I1201 09:19:37.143264 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cf256dda-e402-4a9f-bff1-fe2990f7ce72-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kntht\" (UID: \"cf256dda-e402-4a9f-bff1-fe2990f7ce72\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kntht" Dec 01 09:19:37 crc kubenswrapper[4873]: I1201 09:19:37.245658 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqrnt\" (UniqueName: \"kubernetes.io/projected/cf256dda-e402-4a9f-bff1-fe2990f7ce72-kube-api-access-zqrnt\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kntht\" (UID: \"cf256dda-e402-4a9f-bff1-fe2990f7ce72\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kntht" Dec 01 09:19:37 crc kubenswrapper[4873]: I1201 09:19:37.245724 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cf256dda-e402-4a9f-bff1-fe2990f7ce72-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kntht\" (UID: \"cf256dda-e402-4a9f-bff1-fe2990f7ce72\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kntht" Dec 01 09:19:37 crc kubenswrapper[4873]: I1201 09:19:37.245772 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cf256dda-e402-4a9f-bff1-fe2990f7ce72-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kntht\" (UID: \"cf256dda-e402-4a9f-bff1-fe2990f7ce72\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kntht" Dec 01 09:19:37 crc kubenswrapper[4873]: I1201 09:19:37.245930 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cf256dda-e402-4a9f-bff1-fe2990f7ce72-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kntht\" (UID: \"cf256dda-e402-4a9f-bff1-fe2990f7ce72\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kntht" Dec 01 09:19:37 crc kubenswrapper[4873]: I1201 09:19:37.251102 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cf256dda-e402-4a9f-bff1-fe2990f7ce72-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kntht\" (UID: \"cf256dda-e402-4a9f-bff1-fe2990f7ce72\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kntht" Dec 01 09:19:37 crc kubenswrapper[4873]: I1201 09:19:37.251191 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cf256dda-e402-4a9f-bff1-fe2990f7ce72-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kntht\" (UID: \"cf256dda-e402-4a9f-bff1-fe2990f7ce72\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kntht" Dec 01 09:19:37 crc kubenswrapper[4873]: I1201 09:19:37.253124 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cf256dda-e402-4a9f-bff1-fe2990f7ce72-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kntht\" (UID: \"cf256dda-e402-4a9f-bff1-fe2990f7ce72\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kntht" Dec 01 09:19:37 crc kubenswrapper[4873]: I1201 09:19:37.270937 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqrnt\" (UniqueName: \"kubernetes.io/projected/cf256dda-e402-4a9f-bff1-fe2990f7ce72-kube-api-access-zqrnt\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kntht\" (UID: \"cf256dda-e402-4a9f-bff1-fe2990f7ce72\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kntht" Dec 01 09:19:37 crc kubenswrapper[4873]: I1201 09:19:37.382829 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kntht" Dec 01 09:19:37 crc kubenswrapper[4873]: I1201 09:19:37.739528 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kntht"] Dec 01 09:19:37 crc kubenswrapper[4873]: I1201 09:19:37.959840 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kntht" event={"ID":"cf256dda-e402-4a9f-bff1-fe2990f7ce72","Type":"ContainerStarted","Data":"92484eb51c178fa022f317be44648798605ec3ea98788599bfb7e1c12c3da453"} Dec 01 09:19:38 crc kubenswrapper[4873]: I1201 09:19:38.969726 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kntht" event={"ID":"cf256dda-e402-4a9f-bff1-fe2990f7ce72","Type":"ContainerStarted","Data":"9d4da10c90df373f5bbfe6b7fcf951e50a490b22d0632656526d070ba8860202"} Dec 01 09:19:38 crc kubenswrapper[4873]: I1201 09:19:38.992994 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kntht" podStartSLOduration=1.446404417 podStartE2EDuration="1.992971438s" podCreationTimestamp="2025-12-01 09:19:37 +0000 UTC" firstStartedPulling="2025-12-01 09:19:37.746916498 +0000 UTC m=+2353.649025037" lastFinishedPulling="2025-12-01 09:19:38.293483509 +0000 UTC m=+2354.195592058" observedRunningTime="2025-12-01 09:19:38.98940178 +0000 UTC m=+2354.891510319" watchObservedRunningTime="2025-12-01 09:19:38.992971438 +0000 UTC m=+2354.895079977" Dec 01 09:19:42 crc kubenswrapper[4873]: I1201 09:19:42.429877 4873 scope.go:117] "RemoveContainer" containerID="1a1db8f8327e8cf6dbb94b57ea898b250339bb198ae1ba595cb1d79805fa0db1" Dec 01 09:19:42 crc kubenswrapper[4873]: E1201 09:19:42.430942 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:19:53 crc kubenswrapper[4873]: I1201 09:19:53.429908 4873 scope.go:117] "RemoveContainer" containerID="1a1db8f8327e8cf6dbb94b57ea898b250339bb198ae1ba595cb1d79805fa0db1" Dec 01 09:19:53 crc kubenswrapper[4873]: E1201 09:19:53.430950 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:20:06 crc kubenswrapper[4873]: I1201 09:20:06.430981 4873 scope.go:117] "RemoveContainer" containerID="1a1db8f8327e8cf6dbb94b57ea898b250339bb198ae1ba595cb1d79805fa0db1" Dec 01 09:20:06 crc kubenswrapper[4873]: E1201 09:20:06.432227 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:20:20 crc kubenswrapper[4873]: I1201 09:20:20.430752 4873 scope.go:117] "RemoveContainer" containerID="1a1db8f8327e8cf6dbb94b57ea898b250339bb198ae1ba595cb1d79805fa0db1" Dec 01 09:20:20 crc kubenswrapper[4873]: E1201 09:20:20.431802 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:20:26 crc kubenswrapper[4873]: I1201 09:20:26.446078 4873 generic.go:334] "Generic (PLEG): container finished" podID="cf256dda-e402-4a9f-bff1-fe2990f7ce72" containerID="9d4da10c90df373f5bbfe6b7fcf951e50a490b22d0632656526d070ba8860202" exitCode=0 Dec 01 09:20:26 crc kubenswrapper[4873]: I1201 09:20:26.446179 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kntht" event={"ID":"cf256dda-e402-4a9f-bff1-fe2990f7ce72","Type":"ContainerDied","Data":"9d4da10c90df373f5bbfe6b7fcf951e50a490b22d0632656526d070ba8860202"} Dec 01 09:20:27 crc kubenswrapper[4873]: I1201 09:20:27.890231 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kntht" Dec 01 09:20:28 crc kubenswrapper[4873]: I1201 09:20:28.077471 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cf256dda-e402-4a9f-bff1-fe2990f7ce72-inventory\") pod \"cf256dda-e402-4a9f-bff1-fe2990f7ce72\" (UID: \"cf256dda-e402-4a9f-bff1-fe2990f7ce72\") " Dec 01 09:20:28 crc kubenswrapper[4873]: I1201 09:20:28.077619 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cf256dda-e402-4a9f-bff1-fe2990f7ce72-ssh-key\") pod \"cf256dda-e402-4a9f-bff1-fe2990f7ce72\" (UID: \"cf256dda-e402-4a9f-bff1-fe2990f7ce72\") " Dec 01 09:20:28 crc kubenswrapper[4873]: I1201 09:20:28.077737 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cf256dda-e402-4a9f-bff1-fe2990f7ce72-ceph\") pod \"cf256dda-e402-4a9f-bff1-fe2990f7ce72\" (UID: \"cf256dda-e402-4a9f-bff1-fe2990f7ce72\") " Dec 01 09:20:28 crc kubenswrapper[4873]: I1201 09:20:28.077915 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zqrnt\" (UniqueName: \"kubernetes.io/projected/cf256dda-e402-4a9f-bff1-fe2990f7ce72-kube-api-access-zqrnt\") pod \"cf256dda-e402-4a9f-bff1-fe2990f7ce72\" (UID: \"cf256dda-e402-4a9f-bff1-fe2990f7ce72\") " Dec 01 09:20:28 crc kubenswrapper[4873]: I1201 09:20:28.084891 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf256dda-e402-4a9f-bff1-fe2990f7ce72-ceph" (OuterVolumeSpecName: "ceph") pod "cf256dda-e402-4a9f-bff1-fe2990f7ce72" (UID: "cf256dda-e402-4a9f-bff1-fe2990f7ce72"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:20:28 crc kubenswrapper[4873]: I1201 09:20:28.089382 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf256dda-e402-4a9f-bff1-fe2990f7ce72-kube-api-access-zqrnt" (OuterVolumeSpecName: "kube-api-access-zqrnt") pod "cf256dda-e402-4a9f-bff1-fe2990f7ce72" (UID: "cf256dda-e402-4a9f-bff1-fe2990f7ce72"). InnerVolumeSpecName "kube-api-access-zqrnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:20:28 crc kubenswrapper[4873]: I1201 09:20:28.106901 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf256dda-e402-4a9f-bff1-fe2990f7ce72-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "cf256dda-e402-4a9f-bff1-fe2990f7ce72" (UID: "cf256dda-e402-4a9f-bff1-fe2990f7ce72"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:20:28 crc kubenswrapper[4873]: I1201 09:20:28.113072 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf256dda-e402-4a9f-bff1-fe2990f7ce72-inventory" (OuterVolumeSpecName: "inventory") pod "cf256dda-e402-4a9f-bff1-fe2990f7ce72" (UID: "cf256dda-e402-4a9f-bff1-fe2990f7ce72"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:20:28 crc kubenswrapper[4873]: I1201 09:20:28.180629 4873 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cf256dda-e402-4a9f-bff1-fe2990f7ce72-ceph\") on node \"crc\" DevicePath \"\"" Dec 01 09:20:28 crc kubenswrapper[4873]: I1201 09:20:28.180681 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zqrnt\" (UniqueName: \"kubernetes.io/projected/cf256dda-e402-4a9f-bff1-fe2990f7ce72-kube-api-access-zqrnt\") on node \"crc\" DevicePath \"\"" Dec 01 09:20:28 crc kubenswrapper[4873]: I1201 09:20:28.180699 4873 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cf256dda-e402-4a9f-bff1-fe2990f7ce72-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 09:20:28 crc kubenswrapper[4873]: I1201 09:20:28.180715 4873 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cf256dda-e402-4a9f-bff1-fe2990f7ce72-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 09:20:28 crc kubenswrapper[4873]: I1201 09:20:28.469600 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kntht" event={"ID":"cf256dda-e402-4a9f-bff1-fe2990f7ce72","Type":"ContainerDied","Data":"92484eb51c178fa022f317be44648798605ec3ea98788599bfb7e1c12c3da453"} Dec 01 09:20:28 crc kubenswrapper[4873]: I1201 09:20:28.469647 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="92484eb51c178fa022f317be44648798605ec3ea98788599bfb7e1c12c3da453" Dec 01 09:20:28 crc kubenswrapper[4873]: I1201 09:20:28.469647 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kntht" Dec 01 09:20:28 crc kubenswrapper[4873]: I1201 09:20:28.584388 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-9k6m8"] Dec 01 09:20:28 crc kubenswrapper[4873]: E1201 09:20:28.584827 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf256dda-e402-4a9f-bff1-fe2990f7ce72" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 01 09:20:28 crc kubenswrapper[4873]: I1201 09:20:28.584848 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf256dda-e402-4a9f-bff1-fe2990f7ce72" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 01 09:20:28 crc kubenswrapper[4873]: I1201 09:20:28.585057 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf256dda-e402-4a9f-bff1-fe2990f7ce72" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 01 09:20:28 crc kubenswrapper[4873]: I1201 09:20:28.585702 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-9k6m8" Dec 01 09:20:28 crc kubenswrapper[4873]: I1201 09:20:28.587763 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 01 09:20:28 crc kubenswrapper[4873]: I1201 09:20:28.589500 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 09:20:28 crc kubenswrapper[4873]: I1201 09:20:28.589815 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 09:20:28 crc kubenswrapper[4873]: I1201 09:20:28.590164 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9ghqw" Dec 01 09:20:28 crc kubenswrapper[4873]: I1201 09:20:28.590163 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 09:20:28 crc kubenswrapper[4873]: I1201 09:20:28.602822 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-9k6m8"] Dec 01 09:20:28 crc kubenswrapper[4873]: I1201 09:20:28.693030 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/2ea95238-7e66-46b0-9c20-b6088f9d6737-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-9k6m8\" (UID: \"2ea95238-7e66-46b0-9c20-b6088f9d6737\") " pod="openstack/ssh-known-hosts-edpm-deployment-9k6m8" Dec 01 09:20:28 crc kubenswrapper[4873]: I1201 09:20:28.693137 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2ea95238-7e66-46b0-9c20-b6088f9d6737-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-9k6m8\" (UID: \"2ea95238-7e66-46b0-9c20-b6088f9d6737\") " pod="openstack/ssh-known-hosts-edpm-deployment-9k6m8" Dec 01 09:20:28 crc kubenswrapper[4873]: I1201 09:20:28.693169 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2ea95238-7e66-46b0-9c20-b6088f9d6737-ceph\") pod \"ssh-known-hosts-edpm-deployment-9k6m8\" (UID: \"2ea95238-7e66-46b0-9c20-b6088f9d6737\") " pod="openstack/ssh-known-hosts-edpm-deployment-9k6m8" Dec 01 09:20:28 crc kubenswrapper[4873]: I1201 09:20:28.693273 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4gd5\" (UniqueName: \"kubernetes.io/projected/2ea95238-7e66-46b0-9c20-b6088f9d6737-kube-api-access-d4gd5\") pod \"ssh-known-hosts-edpm-deployment-9k6m8\" (UID: \"2ea95238-7e66-46b0-9c20-b6088f9d6737\") " pod="openstack/ssh-known-hosts-edpm-deployment-9k6m8" Dec 01 09:20:28 crc kubenswrapper[4873]: I1201 09:20:28.795126 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4gd5\" (UniqueName: \"kubernetes.io/projected/2ea95238-7e66-46b0-9c20-b6088f9d6737-kube-api-access-d4gd5\") pod \"ssh-known-hosts-edpm-deployment-9k6m8\" (UID: \"2ea95238-7e66-46b0-9c20-b6088f9d6737\") " pod="openstack/ssh-known-hosts-edpm-deployment-9k6m8" Dec 01 09:20:28 crc kubenswrapper[4873]: I1201 09:20:28.795228 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/2ea95238-7e66-46b0-9c20-b6088f9d6737-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-9k6m8\" (UID: \"2ea95238-7e66-46b0-9c20-b6088f9d6737\") " pod="openstack/ssh-known-hosts-edpm-deployment-9k6m8" Dec 01 09:20:28 crc kubenswrapper[4873]: I1201 09:20:28.795273 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2ea95238-7e66-46b0-9c20-b6088f9d6737-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-9k6m8\" (UID: \"2ea95238-7e66-46b0-9c20-b6088f9d6737\") " pod="openstack/ssh-known-hosts-edpm-deployment-9k6m8" Dec 01 09:20:28 crc kubenswrapper[4873]: I1201 09:20:28.795295 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2ea95238-7e66-46b0-9c20-b6088f9d6737-ceph\") pod \"ssh-known-hosts-edpm-deployment-9k6m8\" (UID: \"2ea95238-7e66-46b0-9c20-b6088f9d6737\") " pod="openstack/ssh-known-hosts-edpm-deployment-9k6m8" Dec 01 09:20:28 crc kubenswrapper[4873]: I1201 09:20:28.800600 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/2ea95238-7e66-46b0-9c20-b6088f9d6737-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-9k6m8\" (UID: \"2ea95238-7e66-46b0-9c20-b6088f9d6737\") " pod="openstack/ssh-known-hosts-edpm-deployment-9k6m8" Dec 01 09:20:28 crc kubenswrapper[4873]: I1201 09:20:28.800709 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2ea95238-7e66-46b0-9c20-b6088f9d6737-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-9k6m8\" (UID: \"2ea95238-7e66-46b0-9c20-b6088f9d6737\") " pod="openstack/ssh-known-hosts-edpm-deployment-9k6m8" Dec 01 09:20:28 crc kubenswrapper[4873]: I1201 09:20:28.801471 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2ea95238-7e66-46b0-9c20-b6088f9d6737-ceph\") pod \"ssh-known-hosts-edpm-deployment-9k6m8\" (UID: \"2ea95238-7e66-46b0-9c20-b6088f9d6737\") " pod="openstack/ssh-known-hosts-edpm-deployment-9k6m8" Dec 01 09:20:28 crc kubenswrapper[4873]: I1201 09:20:28.823419 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4gd5\" (UniqueName: \"kubernetes.io/projected/2ea95238-7e66-46b0-9c20-b6088f9d6737-kube-api-access-d4gd5\") pod \"ssh-known-hosts-edpm-deployment-9k6m8\" (UID: \"2ea95238-7e66-46b0-9c20-b6088f9d6737\") " pod="openstack/ssh-known-hosts-edpm-deployment-9k6m8" Dec 01 09:20:28 crc kubenswrapper[4873]: I1201 09:20:28.906180 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-9k6m8" Dec 01 09:20:29 crc kubenswrapper[4873]: I1201 09:20:29.485423 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-9k6m8"] Dec 01 09:20:30 crc kubenswrapper[4873]: I1201 09:20:30.495052 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-9k6m8" event={"ID":"2ea95238-7e66-46b0-9c20-b6088f9d6737","Type":"ContainerStarted","Data":"ba4b93c696b25e00a680bedc7392bf675bf8cca7040c8dee9559e22ee8bcb9c5"} Dec 01 09:20:31 crc kubenswrapper[4873]: I1201 09:20:31.433669 4873 scope.go:117] "RemoveContainer" containerID="1a1db8f8327e8cf6dbb94b57ea898b250339bb198ae1ba595cb1d79805fa0db1" Dec 01 09:20:31 crc kubenswrapper[4873]: E1201 09:20:31.433978 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:20:31 crc kubenswrapper[4873]: I1201 09:20:31.507708 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-9k6m8" event={"ID":"2ea95238-7e66-46b0-9c20-b6088f9d6737","Type":"ContainerStarted","Data":"54f2156e0537d02f9f20ea6141e325cf3c7f5ed585107265921e839d4f69e32a"} Dec 01 09:20:31 crc kubenswrapper[4873]: I1201 09:20:31.537927 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-9k6m8" podStartSLOduration=2.798209185 podStartE2EDuration="3.537901127s" podCreationTimestamp="2025-12-01 09:20:28 +0000 UTC" firstStartedPulling="2025-12-01 09:20:29.498310796 +0000 UTC m=+2405.400419335" lastFinishedPulling="2025-12-01 09:20:30.238002738 +0000 UTC m=+2406.140111277" observedRunningTime="2025-12-01 09:20:31.52221262 +0000 UTC m=+2407.424321159" watchObservedRunningTime="2025-12-01 09:20:31.537901127 +0000 UTC m=+2407.440009666" Dec 01 09:20:40 crc kubenswrapper[4873]: I1201 09:20:40.616140 4873 generic.go:334] "Generic (PLEG): container finished" podID="2ea95238-7e66-46b0-9c20-b6088f9d6737" containerID="54f2156e0537d02f9f20ea6141e325cf3c7f5ed585107265921e839d4f69e32a" exitCode=0 Dec 01 09:20:40 crc kubenswrapper[4873]: I1201 09:20:40.616200 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-9k6m8" event={"ID":"2ea95238-7e66-46b0-9c20-b6088f9d6737","Type":"ContainerDied","Data":"54f2156e0537d02f9f20ea6141e325cf3c7f5ed585107265921e839d4f69e32a"} Dec 01 09:20:42 crc kubenswrapper[4873]: I1201 09:20:42.107743 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-9k6m8" Dec 01 09:20:42 crc kubenswrapper[4873]: I1201 09:20:42.208438 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/2ea95238-7e66-46b0-9c20-b6088f9d6737-inventory-0\") pod \"2ea95238-7e66-46b0-9c20-b6088f9d6737\" (UID: \"2ea95238-7e66-46b0-9c20-b6088f9d6737\") " Dec 01 09:20:42 crc kubenswrapper[4873]: I1201 09:20:42.208546 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4gd5\" (UniqueName: \"kubernetes.io/projected/2ea95238-7e66-46b0-9c20-b6088f9d6737-kube-api-access-d4gd5\") pod \"2ea95238-7e66-46b0-9c20-b6088f9d6737\" (UID: \"2ea95238-7e66-46b0-9c20-b6088f9d6737\") " Dec 01 09:20:42 crc kubenswrapper[4873]: I1201 09:20:42.208589 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2ea95238-7e66-46b0-9c20-b6088f9d6737-ssh-key-openstack-edpm-ipam\") pod \"2ea95238-7e66-46b0-9c20-b6088f9d6737\" (UID: \"2ea95238-7e66-46b0-9c20-b6088f9d6737\") " Dec 01 09:20:42 crc kubenswrapper[4873]: I1201 09:20:42.208628 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2ea95238-7e66-46b0-9c20-b6088f9d6737-ceph\") pod \"2ea95238-7e66-46b0-9c20-b6088f9d6737\" (UID: \"2ea95238-7e66-46b0-9c20-b6088f9d6737\") " Dec 01 09:20:42 crc kubenswrapper[4873]: I1201 09:20:42.216722 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ea95238-7e66-46b0-9c20-b6088f9d6737-ceph" (OuterVolumeSpecName: "ceph") pod "2ea95238-7e66-46b0-9c20-b6088f9d6737" (UID: "2ea95238-7e66-46b0-9c20-b6088f9d6737"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:20:42 crc kubenswrapper[4873]: I1201 09:20:42.217148 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ea95238-7e66-46b0-9c20-b6088f9d6737-kube-api-access-d4gd5" (OuterVolumeSpecName: "kube-api-access-d4gd5") pod "2ea95238-7e66-46b0-9c20-b6088f9d6737" (UID: "2ea95238-7e66-46b0-9c20-b6088f9d6737"). InnerVolumeSpecName "kube-api-access-d4gd5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:20:42 crc kubenswrapper[4873]: I1201 09:20:42.237590 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ea95238-7e66-46b0-9c20-b6088f9d6737-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "2ea95238-7e66-46b0-9c20-b6088f9d6737" (UID: "2ea95238-7e66-46b0-9c20-b6088f9d6737"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:20:42 crc kubenswrapper[4873]: I1201 09:20:42.242891 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ea95238-7e66-46b0-9c20-b6088f9d6737-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "2ea95238-7e66-46b0-9c20-b6088f9d6737" (UID: "2ea95238-7e66-46b0-9c20-b6088f9d6737"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:20:42 crc kubenswrapper[4873]: I1201 09:20:42.311308 4873 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/2ea95238-7e66-46b0-9c20-b6088f9d6737-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 01 09:20:42 crc kubenswrapper[4873]: I1201 09:20:42.311371 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4gd5\" (UniqueName: \"kubernetes.io/projected/2ea95238-7e66-46b0-9c20-b6088f9d6737-kube-api-access-d4gd5\") on node \"crc\" DevicePath \"\"" Dec 01 09:20:42 crc kubenswrapper[4873]: I1201 09:20:42.311388 4873 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2ea95238-7e66-46b0-9c20-b6088f9d6737-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 01 09:20:42 crc kubenswrapper[4873]: I1201 09:20:42.311403 4873 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2ea95238-7e66-46b0-9c20-b6088f9d6737-ceph\") on node \"crc\" DevicePath \"\"" Dec 01 09:20:42 crc kubenswrapper[4873]: I1201 09:20:42.641234 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-9k6m8" event={"ID":"2ea95238-7e66-46b0-9c20-b6088f9d6737","Type":"ContainerDied","Data":"ba4b93c696b25e00a680bedc7392bf675bf8cca7040c8dee9559e22ee8bcb9c5"} Dec 01 09:20:42 crc kubenswrapper[4873]: I1201 09:20:42.641306 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba4b93c696b25e00a680bedc7392bf675bf8cca7040c8dee9559e22ee8bcb9c5" Dec 01 09:20:42 crc kubenswrapper[4873]: I1201 09:20:42.641371 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-9k6m8" Dec 01 09:20:42 crc kubenswrapper[4873]: I1201 09:20:42.733770 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-gg2dk"] Dec 01 09:20:42 crc kubenswrapper[4873]: E1201 09:20:42.734305 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ea95238-7e66-46b0-9c20-b6088f9d6737" containerName="ssh-known-hosts-edpm-deployment" Dec 01 09:20:42 crc kubenswrapper[4873]: I1201 09:20:42.734329 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ea95238-7e66-46b0-9c20-b6088f9d6737" containerName="ssh-known-hosts-edpm-deployment" Dec 01 09:20:42 crc kubenswrapper[4873]: I1201 09:20:42.734591 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ea95238-7e66-46b0-9c20-b6088f9d6737" containerName="ssh-known-hosts-edpm-deployment" Dec 01 09:20:42 crc kubenswrapper[4873]: I1201 09:20:42.735414 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gg2dk" Dec 01 09:20:42 crc kubenswrapper[4873]: I1201 09:20:42.738160 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 01 09:20:42 crc kubenswrapper[4873]: I1201 09:20:42.738482 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9ghqw" Dec 01 09:20:42 crc kubenswrapper[4873]: I1201 09:20:42.738721 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 09:20:42 crc kubenswrapper[4873]: I1201 09:20:42.738754 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 09:20:42 crc kubenswrapper[4873]: I1201 09:20:42.739648 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 09:20:42 crc kubenswrapper[4873]: I1201 09:20:42.745416 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-gg2dk"] Dec 01 09:20:42 crc kubenswrapper[4873]: I1201 09:20:42.822821 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/74115943-1cf4-48ce-9f8c-41de87f8fe52-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gg2dk\" (UID: \"74115943-1cf4-48ce-9f8c-41de87f8fe52\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gg2dk" Dec 01 09:20:42 crc kubenswrapper[4873]: I1201 09:20:42.822898 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6xjm\" (UniqueName: \"kubernetes.io/projected/74115943-1cf4-48ce-9f8c-41de87f8fe52-kube-api-access-v6xjm\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gg2dk\" (UID: \"74115943-1cf4-48ce-9f8c-41de87f8fe52\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gg2dk" Dec 01 09:20:42 crc kubenswrapper[4873]: I1201 09:20:42.822928 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74115943-1cf4-48ce-9f8c-41de87f8fe52-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gg2dk\" (UID: \"74115943-1cf4-48ce-9f8c-41de87f8fe52\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gg2dk" Dec 01 09:20:42 crc kubenswrapper[4873]: I1201 09:20:42.822967 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/74115943-1cf4-48ce-9f8c-41de87f8fe52-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gg2dk\" (UID: \"74115943-1cf4-48ce-9f8c-41de87f8fe52\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gg2dk" Dec 01 09:20:42 crc kubenswrapper[4873]: I1201 09:20:42.925933 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/74115943-1cf4-48ce-9f8c-41de87f8fe52-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gg2dk\" (UID: \"74115943-1cf4-48ce-9f8c-41de87f8fe52\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gg2dk" Dec 01 09:20:42 crc kubenswrapper[4873]: I1201 09:20:42.926003 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6xjm\" (UniqueName: \"kubernetes.io/projected/74115943-1cf4-48ce-9f8c-41de87f8fe52-kube-api-access-v6xjm\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gg2dk\" (UID: \"74115943-1cf4-48ce-9f8c-41de87f8fe52\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gg2dk" Dec 01 09:20:42 crc kubenswrapper[4873]: I1201 09:20:42.926068 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74115943-1cf4-48ce-9f8c-41de87f8fe52-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gg2dk\" (UID: \"74115943-1cf4-48ce-9f8c-41de87f8fe52\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gg2dk" Dec 01 09:20:42 crc kubenswrapper[4873]: I1201 09:20:42.926112 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/74115943-1cf4-48ce-9f8c-41de87f8fe52-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gg2dk\" (UID: \"74115943-1cf4-48ce-9f8c-41de87f8fe52\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gg2dk" Dec 01 09:20:42 crc kubenswrapper[4873]: I1201 09:20:42.930946 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/74115943-1cf4-48ce-9f8c-41de87f8fe52-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gg2dk\" (UID: \"74115943-1cf4-48ce-9f8c-41de87f8fe52\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gg2dk" Dec 01 09:20:42 crc kubenswrapper[4873]: I1201 09:20:42.931108 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/74115943-1cf4-48ce-9f8c-41de87f8fe52-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gg2dk\" (UID: \"74115943-1cf4-48ce-9f8c-41de87f8fe52\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gg2dk" Dec 01 09:20:42 crc kubenswrapper[4873]: I1201 09:20:42.933294 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74115943-1cf4-48ce-9f8c-41de87f8fe52-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gg2dk\" (UID: \"74115943-1cf4-48ce-9f8c-41de87f8fe52\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gg2dk" Dec 01 09:20:42 crc kubenswrapper[4873]: I1201 09:20:42.951285 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6xjm\" (UniqueName: \"kubernetes.io/projected/74115943-1cf4-48ce-9f8c-41de87f8fe52-kube-api-access-v6xjm\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gg2dk\" (UID: \"74115943-1cf4-48ce-9f8c-41de87f8fe52\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gg2dk" Dec 01 09:20:43 crc kubenswrapper[4873]: I1201 09:20:43.055981 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gg2dk" Dec 01 09:20:43 crc kubenswrapper[4873]: I1201 09:20:43.581308 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-gg2dk"] Dec 01 09:20:43 crc kubenswrapper[4873]: I1201 09:20:43.654137 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gg2dk" event={"ID":"74115943-1cf4-48ce-9f8c-41de87f8fe52","Type":"ContainerStarted","Data":"481e1b84393e5526505e07ff2641a29b31872810fcbcf0fa98bc30163710b300"} Dec 01 09:20:44 crc kubenswrapper[4873]: I1201 09:20:44.666725 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gg2dk" event={"ID":"74115943-1cf4-48ce-9f8c-41de87f8fe52","Type":"ContainerStarted","Data":"c97941284914945bf05c5d0762aa2aeb59538cb18ae54dbf0a3fb9ee5c797f33"} Dec 01 09:20:44 crc kubenswrapper[4873]: I1201 09:20:44.691340 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gg2dk" podStartSLOduration=2.083428535 podStartE2EDuration="2.691318961s" podCreationTimestamp="2025-12-01 09:20:42 +0000 UTC" firstStartedPulling="2025-12-01 09:20:43.590100348 +0000 UTC m=+2419.492208887" lastFinishedPulling="2025-12-01 09:20:44.197990774 +0000 UTC m=+2420.100099313" observedRunningTime="2025-12-01 09:20:44.685443765 +0000 UTC m=+2420.587552304" watchObservedRunningTime="2025-12-01 09:20:44.691318961 +0000 UTC m=+2420.593427500" Dec 01 09:20:45 crc kubenswrapper[4873]: I1201 09:20:45.430459 4873 scope.go:117] "RemoveContainer" containerID="1a1db8f8327e8cf6dbb94b57ea898b250339bb198ae1ba595cb1d79805fa0db1" Dec 01 09:20:45 crc kubenswrapper[4873]: E1201 09:20:45.430850 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:20:52 crc kubenswrapper[4873]: I1201 09:20:52.743557 4873 generic.go:334] "Generic (PLEG): container finished" podID="74115943-1cf4-48ce-9f8c-41de87f8fe52" containerID="c97941284914945bf05c5d0762aa2aeb59538cb18ae54dbf0a3fb9ee5c797f33" exitCode=0 Dec 01 09:20:52 crc kubenswrapper[4873]: I1201 09:20:52.743656 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gg2dk" event={"ID":"74115943-1cf4-48ce-9f8c-41de87f8fe52","Type":"ContainerDied","Data":"c97941284914945bf05c5d0762aa2aeb59538cb18ae54dbf0a3fb9ee5c797f33"} Dec 01 09:20:54 crc kubenswrapper[4873]: I1201 09:20:54.215328 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gg2dk" Dec 01 09:20:54 crc kubenswrapper[4873]: I1201 09:20:54.411854 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6xjm\" (UniqueName: \"kubernetes.io/projected/74115943-1cf4-48ce-9f8c-41de87f8fe52-kube-api-access-v6xjm\") pod \"74115943-1cf4-48ce-9f8c-41de87f8fe52\" (UID: \"74115943-1cf4-48ce-9f8c-41de87f8fe52\") " Dec 01 09:20:54 crc kubenswrapper[4873]: I1201 09:20:54.412220 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/74115943-1cf4-48ce-9f8c-41de87f8fe52-ceph\") pod \"74115943-1cf4-48ce-9f8c-41de87f8fe52\" (UID: \"74115943-1cf4-48ce-9f8c-41de87f8fe52\") " Dec 01 09:20:54 crc kubenswrapper[4873]: I1201 09:20:54.412291 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74115943-1cf4-48ce-9f8c-41de87f8fe52-ssh-key\") pod \"74115943-1cf4-48ce-9f8c-41de87f8fe52\" (UID: \"74115943-1cf4-48ce-9f8c-41de87f8fe52\") " Dec 01 09:20:54 crc kubenswrapper[4873]: I1201 09:20:54.412342 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/74115943-1cf4-48ce-9f8c-41de87f8fe52-inventory\") pod \"74115943-1cf4-48ce-9f8c-41de87f8fe52\" (UID: \"74115943-1cf4-48ce-9f8c-41de87f8fe52\") " Dec 01 09:20:54 crc kubenswrapper[4873]: I1201 09:20:54.420676 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74115943-1cf4-48ce-9f8c-41de87f8fe52-kube-api-access-v6xjm" (OuterVolumeSpecName: "kube-api-access-v6xjm") pod "74115943-1cf4-48ce-9f8c-41de87f8fe52" (UID: "74115943-1cf4-48ce-9f8c-41de87f8fe52"). InnerVolumeSpecName "kube-api-access-v6xjm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:20:54 crc kubenswrapper[4873]: I1201 09:20:54.420780 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74115943-1cf4-48ce-9f8c-41de87f8fe52-ceph" (OuterVolumeSpecName: "ceph") pod "74115943-1cf4-48ce-9f8c-41de87f8fe52" (UID: "74115943-1cf4-48ce-9f8c-41de87f8fe52"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:20:54 crc kubenswrapper[4873]: I1201 09:20:54.447751 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74115943-1cf4-48ce-9f8c-41de87f8fe52-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "74115943-1cf4-48ce-9f8c-41de87f8fe52" (UID: "74115943-1cf4-48ce-9f8c-41de87f8fe52"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:20:54 crc kubenswrapper[4873]: I1201 09:20:54.457816 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74115943-1cf4-48ce-9f8c-41de87f8fe52-inventory" (OuterVolumeSpecName: "inventory") pod "74115943-1cf4-48ce-9f8c-41de87f8fe52" (UID: "74115943-1cf4-48ce-9f8c-41de87f8fe52"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:20:54 crc kubenswrapper[4873]: I1201 09:20:54.515041 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6xjm\" (UniqueName: \"kubernetes.io/projected/74115943-1cf4-48ce-9f8c-41de87f8fe52-kube-api-access-v6xjm\") on node \"crc\" DevicePath \"\"" Dec 01 09:20:54 crc kubenswrapper[4873]: I1201 09:20:54.515088 4873 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/74115943-1cf4-48ce-9f8c-41de87f8fe52-ceph\") on node \"crc\" DevicePath \"\"" Dec 01 09:20:54 crc kubenswrapper[4873]: I1201 09:20:54.515098 4873 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74115943-1cf4-48ce-9f8c-41de87f8fe52-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 09:20:54 crc kubenswrapper[4873]: I1201 09:20:54.515106 4873 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/74115943-1cf4-48ce-9f8c-41de87f8fe52-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 09:20:54 crc kubenswrapper[4873]: I1201 09:20:54.765076 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gg2dk" event={"ID":"74115943-1cf4-48ce-9f8c-41de87f8fe52","Type":"ContainerDied","Data":"481e1b84393e5526505e07ff2641a29b31872810fcbcf0fa98bc30163710b300"} Dec 01 09:20:54 crc kubenswrapper[4873]: I1201 09:20:54.765140 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="481e1b84393e5526505e07ff2641a29b31872810fcbcf0fa98bc30163710b300" Dec 01 09:20:54 crc kubenswrapper[4873]: I1201 09:20:54.765174 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gg2dk" Dec 01 09:20:54 crc kubenswrapper[4873]: I1201 09:20:54.851991 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9bdgf"] Dec 01 09:20:54 crc kubenswrapper[4873]: E1201 09:20:54.852633 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74115943-1cf4-48ce-9f8c-41de87f8fe52" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 01 09:20:54 crc kubenswrapper[4873]: I1201 09:20:54.852664 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="74115943-1cf4-48ce-9f8c-41de87f8fe52" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 01 09:20:54 crc kubenswrapper[4873]: I1201 09:20:54.852920 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="74115943-1cf4-48ce-9f8c-41de87f8fe52" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 01 09:20:54 crc kubenswrapper[4873]: I1201 09:20:54.856775 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9bdgf" Dec 01 09:20:54 crc kubenswrapper[4873]: I1201 09:20:54.864401 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 09:20:54 crc kubenswrapper[4873]: I1201 09:20:54.864513 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 01 09:20:54 crc kubenswrapper[4873]: I1201 09:20:54.864677 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9ghqw" Dec 01 09:20:54 crc kubenswrapper[4873]: I1201 09:20:54.864401 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 09:20:54 crc kubenswrapper[4873]: I1201 09:20:54.864444 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 09:20:54 crc kubenswrapper[4873]: I1201 09:20:54.867058 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9bdgf"] Dec 01 09:20:55 crc kubenswrapper[4873]: I1201 09:20:55.024973 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e301f3ff-9983-4440-ad43-96ef975437f7-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-9bdgf\" (UID: \"e301f3ff-9983-4440-ad43-96ef975437f7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9bdgf" Dec 01 09:20:55 crc kubenswrapper[4873]: I1201 09:20:55.025170 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e301f3ff-9983-4440-ad43-96ef975437f7-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-9bdgf\" (UID: \"e301f3ff-9983-4440-ad43-96ef975437f7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9bdgf" Dec 01 09:20:55 crc kubenswrapper[4873]: I1201 09:20:55.025277 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e301f3ff-9983-4440-ad43-96ef975437f7-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-9bdgf\" (UID: \"e301f3ff-9983-4440-ad43-96ef975437f7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9bdgf" Dec 01 09:20:55 crc kubenswrapper[4873]: I1201 09:20:55.025391 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtqkf\" (UniqueName: \"kubernetes.io/projected/e301f3ff-9983-4440-ad43-96ef975437f7-kube-api-access-qtqkf\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-9bdgf\" (UID: \"e301f3ff-9983-4440-ad43-96ef975437f7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9bdgf" Dec 01 09:20:55 crc kubenswrapper[4873]: I1201 09:20:55.127578 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e301f3ff-9983-4440-ad43-96ef975437f7-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-9bdgf\" (UID: \"e301f3ff-9983-4440-ad43-96ef975437f7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9bdgf" Dec 01 09:20:55 crc kubenswrapper[4873]: I1201 09:20:55.127670 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e301f3ff-9983-4440-ad43-96ef975437f7-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-9bdgf\" (UID: \"e301f3ff-9983-4440-ad43-96ef975437f7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9bdgf" Dec 01 09:20:55 crc kubenswrapper[4873]: I1201 09:20:55.127698 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e301f3ff-9983-4440-ad43-96ef975437f7-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-9bdgf\" (UID: \"e301f3ff-9983-4440-ad43-96ef975437f7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9bdgf" Dec 01 09:20:55 crc kubenswrapper[4873]: I1201 09:20:55.127729 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtqkf\" (UniqueName: \"kubernetes.io/projected/e301f3ff-9983-4440-ad43-96ef975437f7-kube-api-access-qtqkf\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-9bdgf\" (UID: \"e301f3ff-9983-4440-ad43-96ef975437f7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9bdgf" Dec 01 09:20:55 crc kubenswrapper[4873]: I1201 09:20:55.143307 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e301f3ff-9983-4440-ad43-96ef975437f7-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-9bdgf\" (UID: \"e301f3ff-9983-4440-ad43-96ef975437f7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9bdgf" Dec 01 09:20:55 crc kubenswrapper[4873]: I1201 09:20:55.143536 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e301f3ff-9983-4440-ad43-96ef975437f7-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-9bdgf\" (UID: \"e301f3ff-9983-4440-ad43-96ef975437f7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9bdgf" Dec 01 09:20:55 crc kubenswrapper[4873]: I1201 09:20:55.143539 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e301f3ff-9983-4440-ad43-96ef975437f7-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-9bdgf\" (UID: \"e301f3ff-9983-4440-ad43-96ef975437f7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9bdgf" Dec 01 09:20:55 crc kubenswrapper[4873]: I1201 09:20:55.148642 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtqkf\" (UniqueName: \"kubernetes.io/projected/e301f3ff-9983-4440-ad43-96ef975437f7-kube-api-access-qtqkf\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-9bdgf\" (UID: \"e301f3ff-9983-4440-ad43-96ef975437f7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9bdgf" Dec 01 09:20:55 crc kubenswrapper[4873]: I1201 09:20:55.183874 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9bdgf" Dec 01 09:20:55 crc kubenswrapper[4873]: I1201 09:20:55.809381 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9bdgf"] Dec 01 09:20:56 crc kubenswrapper[4873]: I1201 09:20:56.785396 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9bdgf" event={"ID":"e301f3ff-9983-4440-ad43-96ef975437f7","Type":"ContainerStarted","Data":"27fb1900f37eebe59a8fef514f7530cd9c98c27ffd2bb65b146af952ecb6d595"} Dec 01 09:20:57 crc kubenswrapper[4873]: I1201 09:20:57.806469 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9bdgf" event={"ID":"e301f3ff-9983-4440-ad43-96ef975437f7","Type":"ContainerStarted","Data":"e5da886bf656e1348e1ea2e9f35f8f26b1058d05dbd0e6898cbc59226d10e275"} Dec 01 09:20:57 crc kubenswrapper[4873]: I1201 09:20:57.833047 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9bdgf" podStartSLOduration=2.997827304 podStartE2EDuration="3.833026724s" podCreationTimestamp="2025-12-01 09:20:54 +0000 UTC" firstStartedPulling="2025-12-01 09:20:55.812768211 +0000 UTC m=+2431.714876750" lastFinishedPulling="2025-12-01 09:20:56.647967631 +0000 UTC m=+2432.550076170" observedRunningTime="2025-12-01 09:20:57.827709303 +0000 UTC m=+2433.729817842" watchObservedRunningTime="2025-12-01 09:20:57.833026724 +0000 UTC m=+2433.735135263" Dec 01 09:21:00 crc kubenswrapper[4873]: I1201 09:21:00.431176 4873 scope.go:117] "RemoveContainer" containerID="1a1db8f8327e8cf6dbb94b57ea898b250339bb198ae1ba595cb1d79805fa0db1" Dec 01 09:21:00 crc kubenswrapper[4873]: E1201 09:21:00.431761 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:21:07 crc kubenswrapper[4873]: I1201 09:21:07.903311 4873 generic.go:334] "Generic (PLEG): container finished" podID="e301f3ff-9983-4440-ad43-96ef975437f7" containerID="e5da886bf656e1348e1ea2e9f35f8f26b1058d05dbd0e6898cbc59226d10e275" exitCode=0 Dec 01 09:21:07 crc kubenswrapper[4873]: I1201 09:21:07.903395 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9bdgf" event={"ID":"e301f3ff-9983-4440-ad43-96ef975437f7","Type":"ContainerDied","Data":"e5da886bf656e1348e1ea2e9f35f8f26b1058d05dbd0e6898cbc59226d10e275"} Dec 01 09:21:09 crc kubenswrapper[4873]: I1201 09:21:09.360844 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9bdgf" Dec 01 09:21:09 crc kubenswrapper[4873]: I1201 09:21:09.444208 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e301f3ff-9983-4440-ad43-96ef975437f7-inventory\") pod \"e301f3ff-9983-4440-ad43-96ef975437f7\" (UID: \"e301f3ff-9983-4440-ad43-96ef975437f7\") " Dec 01 09:21:09 crc kubenswrapper[4873]: I1201 09:21:09.444342 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qtqkf\" (UniqueName: \"kubernetes.io/projected/e301f3ff-9983-4440-ad43-96ef975437f7-kube-api-access-qtqkf\") pod \"e301f3ff-9983-4440-ad43-96ef975437f7\" (UID: \"e301f3ff-9983-4440-ad43-96ef975437f7\") " Dec 01 09:21:09 crc kubenswrapper[4873]: I1201 09:21:09.444478 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e301f3ff-9983-4440-ad43-96ef975437f7-ssh-key\") pod \"e301f3ff-9983-4440-ad43-96ef975437f7\" (UID: \"e301f3ff-9983-4440-ad43-96ef975437f7\") " Dec 01 09:21:09 crc kubenswrapper[4873]: I1201 09:21:09.444682 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e301f3ff-9983-4440-ad43-96ef975437f7-ceph\") pod \"e301f3ff-9983-4440-ad43-96ef975437f7\" (UID: \"e301f3ff-9983-4440-ad43-96ef975437f7\") " Dec 01 09:21:09 crc kubenswrapper[4873]: I1201 09:21:09.451674 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e301f3ff-9983-4440-ad43-96ef975437f7-kube-api-access-qtqkf" (OuterVolumeSpecName: "kube-api-access-qtqkf") pod "e301f3ff-9983-4440-ad43-96ef975437f7" (UID: "e301f3ff-9983-4440-ad43-96ef975437f7"). InnerVolumeSpecName "kube-api-access-qtqkf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:21:09 crc kubenswrapper[4873]: I1201 09:21:09.452227 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e301f3ff-9983-4440-ad43-96ef975437f7-ceph" (OuterVolumeSpecName: "ceph") pod "e301f3ff-9983-4440-ad43-96ef975437f7" (UID: "e301f3ff-9983-4440-ad43-96ef975437f7"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:21:09 crc kubenswrapper[4873]: I1201 09:21:09.475685 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e301f3ff-9983-4440-ad43-96ef975437f7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e301f3ff-9983-4440-ad43-96ef975437f7" (UID: "e301f3ff-9983-4440-ad43-96ef975437f7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:21:09 crc kubenswrapper[4873]: I1201 09:21:09.478163 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e301f3ff-9983-4440-ad43-96ef975437f7-inventory" (OuterVolumeSpecName: "inventory") pod "e301f3ff-9983-4440-ad43-96ef975437f7" (UID: "e301f3ff-9983-4440-ad43-96ef975437f7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:21:09 crc kubenswrapper[4873]: I1201 09:21:09.547829 4873 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e301f3ff-9983-4440-ad43-96ef975437f7-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 09:21:09 crc kubenswrapper[4873]: I1201 09:21:09.547879 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qtqkf\" (UniqueName: \"kubernetes.io/projected/e301f3ff-9983-4440-ad43-96ef975437f7-kube-api-access-qtqkf\") on node \"crc\" DevicePath \"\"" Dec 01 09:21:09 crc kubenswrapper[4873]: I1201 09:21:09.547897 4873 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e301f3ff-9983-4440-ad43-96ef975437f7-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 09:21:09 crc kubenswrapper[4873]: I1201 09:21:09.547905 4873 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e301f3ff-9983-4440-ad43-96ef975437f7-ceph\") on node \"crc\" DevicePath \"\"" Dec 01 09:21:09 crc kubenswrapper[4873]: I1201 09:21:09.928405 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9bdgf" event={"ID":"e301f3ff-9983-4440-ad43-96ef975437f7","Type":"ContainerDied","Data":"27fb1900f37eebe59a8fef514f7530cd9c98c27ffd2bb65b146af952ecb6d595"} Dec 01 09:21:09 crc kubenswrapper[4873]: I1201 09:21:09.928454 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9bdgf" Dec 01 09:21:09 crc kubenswrapper[4873]: I1201 09:21:09.928469 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="27fb1900f37eebe59a8fef514f7530cd9c98c27ffd2bb65b146af952ecb6d595" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.023037 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj"] Dec 01 09:21:10 crc kubenswrapper[4873]: E1201 09:21:10.023452 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e301f3ff-9983-4440-ad43-96ef975437f7" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.023482 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="e301f3ff-9983-4440-ad43-96ef975437f7" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.023729 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="e301f3ff-9983-4440-ad43-96ef975437f7" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.024676 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.029562 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.029914 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.030256 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.030503 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.030690 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.030930 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.041765 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.041853 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9ghqw" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.046707 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj"] Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.160782 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.160961 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.161217 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/36c9f7db-7138-48d1-9e4d-f04be409c123-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.161424 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.161499 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.161630 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.161687 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvx5p\" (UniqueName: \"kubernetes.io/projected/36c9f7db-7138-48d1-9e4d-f04be409c123-kube-api-access-bvx5p\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.161765 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.161832 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/36c9f7db-7138-48d1-9e4d-f04be409c123-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.161876 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.162158 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/36c9f7db-7138-48d1-9e4d-f04be409c123-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.162289 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.162386 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.265043 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/36c9f7db-7138-48d1-9e4d-f04be409c123-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.265127 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.265175 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.265234 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.265274 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.265320 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/36c9f7db-7138-48d1-9e4d-f04be409c123-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.265375 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.265410 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.265452 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.265484 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvx5p\" (UniqueName: \"kubernetes.io/projected/36c9f7db-7138-48d1-9e4d-f04be409c123-kube-api-access-bvx5p\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.265539 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.265577 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/36c9f7db-7138-48d1-9e4d-f04be409c123-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.265604 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.272242 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.272374 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/36c9f7db-7138-48d1-9e4d-f04be409c123-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.274604 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.274600 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.275323 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.276200 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.276680 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.277224 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.277674 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/36c9f7db-7138-48d1-9e4d-f04be409c123-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.277689 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.282777 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/36c9f7db-7138-48d1-9e4d-f04be409c123-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.286948 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.292516 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvx5p\" (UniqueName: \"kubernetes.io/projected/36c9f7db-7138-48d1-9e4d-f04be409c123-kube-api-access-bvx5p\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.358487 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.906299 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj"] Dec 01 09:21:10 crc kubenswrapper[4873]: W1201 09:21:10.913307 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod36c9f7db_7138_48d1_9e4d_f04be409c123.slice/crio-ed4de16f3d9da1e6af3d58344161fcd61d4a42eca640a1543d7ed3518bcd70de WatchSource:0}: Error finding container ed4de16f3d9da1e6af3d58344161fcd61d4a42eca640a1543d7ed3518bcd70de: Status 404 returned error can't find the container with id ed4de16f3d9da1e6af3d58344161fcd61d4a42eca640a1543d7ed3518bcd70de Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.916301 4873 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 09:21:10 crc kubenswrapper[4873]: I1201 09:21:10.941048 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" event={"ID":"36c9f7db-7138-48d1-9e4d-f04be409c123","Type":"ContainerStarted","Data":"ed4de16f3d9da1e6af3d58344161fcd61d4a42eca640a1543d7ed3518bcd70de"} Dec 01 09:21:11 crc kubenswrapper[4873]: I1201 09:21:11.952711 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" event={"ID":"36c9f7db-7138-48d1-9e4d-f04be409c123","Type":"ContainerStarted","Data":"44f1c094ceb5f4eb006b8f3d986eb1c15b69f20d54dfacbeff2fe153d8ed03bb"} Dec 01 09:21:11 crc kubenswrapper[4873]: I1201 09:21:11.981805 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" podStartSLOduration=2.484658545 podStartE2EDuration="2.981780685s" podCreationTimestamp="2025-12-01 09:21:09 +0000 UTC" firstStartedPulling="2025-12-01 09:21:10.916022788 +0000 UTC m=+2446.818131327" lastFinishedPulling="2025-12-01 09:21:11.413144928 +0000 UTC m=+2447.315253467" observedRunningTime="2025-12-01 09:21:11.971425379 +0000 UTC m=+2447.873533918" watchObservedRunningTime="2025-12-01 09:21:11.981780685 +0000 UTC m=+2447.883889224" Dec 01 09:21:15 crc kubenswrapper[4873]: I1201 09:21:15.432091 4873 scope.go:117] "RemoveContainer" containerID="1a1db8f8327e8cf6dbb94b57ea898b250339bb198ae1ba595cb1d79805fa0db1" Dec 01 09:21:15 crc kubenswrapper[4873]: E1201 09:21:15.432720 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:21:28 crc kubenswrapper[4873]: I1201 09:21:28.431291 4873 scope.go:117] "RemoveContainer" containerID="1a1db8f8327e8cf6dbb94b57ea898b250339bb198ae1ba595cb1d79805fa0db1" Dec 01 09:21:28 crc kubenswrapper[4873]: E1201 09:21:28.432445 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:21:40 crc kubenswrapper[4873]: I1201 09:21:40.431594 4873 scope.go:117] "RemoveContainer" containerID="1a1db8f8327e8cf6dbb94b57ea898b250339bb198ae1ba595cb1d79805fa0db1" Dec 01 09:21:40 crc kubenswrapper[4873]: E1201 09:21:40.432532 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:21:47 crc kubenswrapper[4873]: I1201 09:21:47.335505 4873 generic.go:334] "Generic (PLEG): container finished" podID="36c9f7db-7138-48d1-9e4d-f04be409c123" containerID="44f1c094ceb5f4eb006b8f3d986eb1c15b69f20d54dfacbeff2fe153d8ed03bb" exitCode=0 Dec 01 09:21:47 crc kubenswrapper[4873]: I1201 09:21:47.335898 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" event={"ID":"36c9f7db-7138-48d1-9e4d-f04be409c123","Type":"ContainerDied","Data":"44f1c094ceb5f4eb006b8f3d986eb1c15b69f20d54dfacbeff2fe153d8ed03bb"} Dec 01 09:21:48 crc kubenswrapper[4873]: I1201 09:21:48.814937 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" Dec 01 09:21:48 crc kubenswrapper[4873]: I1201 09:21:48.929445 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-nova-combined-ca-bundle\") pod \"36c9f7db-7138-48d1-9e4d-f04be409c123\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " Dec 01 09:21:48 crc kubenswrapper[4873]: I1201 09:21:48.929511 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/36c9f7db-7138-48d1-9e4d-f04be409c123-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"36c9f7db-7138-48d1-9e4d-f04be409c123\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " Dec 01 09:21:48 crc kubenswrapper[4873]: I1201 09:21:48.929613 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-ssh-key\") pod \"36c9f7db-7138-48d1-9e4d-f04be409c123\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " Dec 01 09:21:48 crc kubenswrapper[4873]: I1201 09:21:48.929666 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-neutron-metadata-combined-ca-bundle\") pod \"36c9f7db-7138-48d1-9e4d-f04be409c123\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " Dec 01 09:21:48 crc kubenswrapper[4873]: I1201 09:21:48.929696 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bvx5p\" (UniqueName: \"kubernetes.io/projected/36c9f7db-7138-48d1-9e4d-f04be409c123-kube-api-access-bvx5p\") pod \"36c9f7db-7138-48d1-9e4d-f04be409c123\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " Dec 01 09:21:48 crc kubenswrapper[4873]: I1201 09:21:48.929773 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-libvirt-combined-ca-bundle\") pod \"36c9f7db-7138-48d1-9e4d-f04be409c123\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " Dec 01 09:21:48 crc kubenswrapper[4873]: I1201 09:21:48.929803 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-bootstrap-combined-ca-bundle\") pod \"36c9f7db-7138-48d1-9e4d-f04be409c123\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " Dec 01 09:21:48 crc kubenswrapper[4873]: I1201 09:21:48.929861 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-inventory\") pod \"36c9f7db-7138-48d1-9e4d-f04be409c123\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " Dec 01 09:21:48 crc kubenswrapper[4873]: I1201 09:21:48.929945 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/36c9f7db-7138-48d1-9e4d-f04be409c123-openstack-edpm-ipam-ovn-default-certs-0\") pod \"36c9f7db-7138-48d1-9e4d-f04be409c123\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " Dec 01 09:21:48 crc kubenswrapper[4873]: I1201 09:21:48.930015 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-repo-setup-combined-ca-bundle\") pod \"36c9f7db-7138-48d1-9e4d-f04be409c123\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " Dec 01 09:21:48 crc kubenswrapper[4873]: I1201 09:21:48.930058 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-ovn-combined-ca-bundle\") pod \"36c9f7db-7138-48d1-9e4d-f04be409c123\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " Dec 01 09:21:48 crc kubenswrapper[4873]: I1201 09:21:48.930084 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-ceph\") pod \"36c9f7db-7138-48d1-9e4d-f04be409c123\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " Dec 01 09:21:48 crc kubenswrapper[4873]: I1201 09:21:48.930145 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/36c9f7db-7138-48d1-9e4d-f04be409c123-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"36c9f7db-7138-48d1-9e4d-f04be409c123\" (UID: \"36c9f7db-7138-48d1-9e4d-f04be409c123\") " Dec 01 09:21:48 crc kubenswrapper[4873]: I1201 09:21:48.938366 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36c9f7db-7138-48d1-9e4d-f04be409c123-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "36c9f7db-7138-48d1-9e4d-f04be409c123" (UID: "36c9f7db-7138-48d1-9e4d-f04be409c123"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:21:48 crc kubenswrapper[4873]: I1201 09:21:48.938784 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-ceph" (OuterVolumeSpecName: "ceph") pod "36c9f7db-7138-48d1-9e4d-f04be409c123" (UID: "36c9f7db-7138-48d1-9e4d-f04be409c123"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:21:48 crc kubenswrapper[4873]: I1201 09:21:48.939074 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36c9f7db-7138-48d1-9e4d-f04be409c123-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "36c9f7db-7138-48d1-9e4d-f04be409c123" (UID: "36c9f7db-7138-48d1-9e4d-f04be409c123"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:21:48 crc kubenswrapper[4873]: I1201 09:21:48.939588 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36c9f7db-7138-48d1-9e4d-f04be409c123-kube-api-access-bvx5p" (OuterVolumeSpecName: "kube-api-access-bvx5p") pod "36c9f7db-7138-48d1-9e4d-f04be409c123" (UID: "36c9f7db-7138-48d1-9e4d-f04be409c123"). InnerVolumeSpecName "kube-api-access-bvx5p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:21:48 crc kubenswrapper[4873]: I1201 09:21:48.939624 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36c9f7db-7138-48d1-9e4d-f04be409c123-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "36c9f7db-7138-48d1-9e4d-f04be409c123" (UID: "36c9f7db-7138-48d1-9e4d-f04be409c123"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:21:48 crc kubenswrapper[4873]: I1201 09:21:48.941052 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "36c9f7db-7138-48d1-9e4d-f04be409c123" (UID: "36c9f7db-7138-48d1-9e4d-f04be409c123"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:21:48 crc kubenswrapper[4873]: I1201 09:21:48.940984 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "36c9f7db-7138-48d1-9e4d-f04be409c123" (UID: "36c9f7db-7138-48d1-9e4d-f04be409c123"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:21:48 crc kubenswrapper[4873]: I1201 09:21:48.941177 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "36c9f7db-7138-48d1-9e4d-f04be409c123" (UID: "36c9f7db-7138-48d1-9e4d-f04be409c123"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:21:48 crc kubenswrapper[4873]: I1201 09:21:48.942385 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "36c9f7db-7138-48d1-9e4d-f04be409c123" (UID: "36c9f7db-7138-48d1-9e4d-f04be409c123"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:21:48 crc kubenswrapper[4873]: I1201 09:21:48.950455 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "36c9f7db-7138-48d1-9e4d-f04be409c123" (UID: "36c9f7db-7138-48d1-9e4d-f04be409c123"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:21:48 crc kubenswrapper[4873]: I1201 09:21:48.950510 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "36c9f7db-7138-48d1-9e4d-f04be409c123" (UID: "36c9f7db-7138-48d1-9e4d-f04be409c123"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:21:48 crc kubenswrapper[4873]: I1201 09:21:48.976255 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "36c9f7db-7138-48d1-9e4d-f04be409c123" (UID: "36c9f7db-7138-48d1-9e4d-f04be409c123"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:21:48 crc kubenswrapper[4873]: I1201 09:21:48.978104 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-inventory" (OuterVolumeSpecName: "inventory") pod "36c9f7db-7138-48d1-9e4d-f04be409c123" (UID: "36c9f7db-7138-48d1-9e4d-f04be409c123"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:21:49 crc kubenswrapper[4873]: I1201 09:21:49.033865 4873 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 09:21:49 crc kubenswrapper[4873]: I1201 09:21:49.033932 4873 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:21:49 crc kubenswrapper[4873]: I1201 09:21:49.033953 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bvx5p\" (UniqueName: \"kubernetes.io/projected/36c9f7db-7138-48d1-9e4d-f04be409c123-kube-api-access-bvx5p\") on node \"crc\" DevicePath \"\"" Dec 01 09:21:49 crc kubenswrapper[4873]: I1201 09:21:49.033969 4873 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:21:49 crc kubenswrapper[4873]: I1201 09:21:49.033982 4873 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:21:49 crc kubenswrapper[4873]: I1201 09:21:49.033996 4873 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 09:21:49 crc kubenswrapper[4873]: I1201 09:21:49.034012 4873 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/36c9f7db-7138-48d1-9e4d-f04be409c123-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 01 09:21:49 crc kubenswrapper[4873]: I1201 09:21:49.034046 4873 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:21:49 crc kubenswrapper[4873]: I1201 09:21:49.034061 4873 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:21:49 crc kubenswrapper[4873]: I1201 09:21:49.034073 4873 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-ceph\") on node \"crc\" DevicePath \"\"" Dec 01 09:21:49 crc kubenswrapper[4873]: I1201 09:21:49.034086 4873 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/36c9f7db-7138-48d1-9e4d-f04be409c123-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 01 09:21:49 crc kubenswrapper[4873]: I1201 09:21:49.034100 4873 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c9f7db-7138-48d1-9e4d-f04be409c123-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:21:49 crc kubenswrapper[4873]: I1201 09:21:49.034113 4873 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/36c9f7db-7138-48d1-9e4d-f04be409c123-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 01 09:21:49 crc kubenswrapper[4873]: I1201 09:21:49.359873 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" event={"ID":"36c9f7db-7138-48d1-9e4d-f04be409c123","Type":"ContainerDied","Data":"ed4de16f3d9da1e6af3d58344161fcd61d4a42eca640a1543d7ed3518bcd70de"} Dec 01 09:21:49 crc kubenswrapper[4873]: I1201 09:21:49.359918 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj" Dec 01 09:21:49 crc kubenswrapper[4873]: I1201 09:21:49.359930 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed4de16f3d9da1e6af3d58344161fcd61d4a42eca640a1543d7ed3518bcd70de" Dec 01 09:21:49 crc kubenswrapper[4873]: I1201 09:21:49.497747 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-cz28q"] Dec 01 09:21:49 crc kubenswrapper[4873]: E1201 09:21:49.498398 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36c9f7db-7138-48d1-9e4d-f04be409c123" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 01 09:21:49 crc kubenswrapper[4873]: I1201 09:21:49.498422 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="36c9f7db-7138-48d1-9e4d-f04be409c123" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 01 09:21:49 crc kubenswrapper[4873]: I1201 09:21:49.498637 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="36c9f7db-7138-48d1-9e4d-f04be409c123" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 01 09:21:49 crc kubenswrapper[4873]: I1201 09:21:49.499548 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-cz28q" Dec 01 09:21:49 crc kubenswrapper[4873]: I1201 09:21:49.504606 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 09:21:49 crc kubenswrapper[4873]: I1201 09:21:49.505132 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 09:21:49 crc kubenswrapper[4873]: I1201 09:21:49.505373 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9ghqw" Dec 01 09:21:49 crc kubenswrapper[4873]: I1201 09:21:49.505592 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 01 09:21:49 crc kubenswrapper[4873]: I1201 09:21:49.505951 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 09:21:49 crc kubenswrapper[4873]: I1201 09:21:49.519981 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-cz28q"] Dec 01 09:21:49 crc kubenswrapper[4873]: I1201 09:21:49.545155 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d186292e-15d4-4944-b6ac-7db1fb7b85f5-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-cz28q\" (UID: \"d186292e-15d4-4944-b6ac-7db1fb7b85f5\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-cz28q" Dec 01 09:21:49 crc kubenswrapper[4873]: I1201 09:21:49.545299 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d186292e-15d4-4944-b6ac-7db1fb7b85f5-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-cz28q\" (UID: \"d186292e-15d4-4944-b6ac-7db1fb7b85f5\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-cz28q" Dec 01 09:21:49 crc kubenswrapper[4873]: I1201 09:21:49.545381 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hll27\" (UniqueName: \"kubernetes.io/projected/d186292e-15d4-4944-b6ac-7db1fb7b85f5-kube-api-access-hll27\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-cz28q\" (UID: \"d186292e-15d4-4944-b6ac-7db1fb7b85f5\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-cz28q" Dec 01 09:21:49 crc kubenswrapper[4873]: I1201 09:21:49.545434 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d186292e-15d4-4944-b6ac-7db1fb7b85f5-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-cz28q\" (UID: \"d186292e-15d4-4944-b6ac-7db1fb7b85f5\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-cz28q" Dec 01 09:21:49 crc kubenswrapper[4873]: I1201 09:21:49.647762 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d186292e-15d4-4944-b6ac-7db1fb7b85f5-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-cz28q\" (UID: \"d186292e-15d4-4944-b6ac-7db1fb7b85f5\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-cz28q" Dec 01 09:21:49 crc kubenswrapper[4873]: I1201 09:21:49.648456 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d186292e-15d4-4944-b6ac-7db1fb7b85f5-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-cz28q\" (UID: \"d186292e-15d4-4944-b6ac-7db1fb7b85f5\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-cz28q" Dec 01 09:21:49 crc kubenswrapper[4873]: I1201 09:21:49.648608 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d186292e-15d4-4944-b6ac-7db1fb7b85f5-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-cz28q\" (UID: \"d186292e-15d4-4944-b6ac-7db1fb7b85f5\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-cz28q" Dec 01 09:21:49 crc kubenswrapper[4873]: I1201 09:21:49.648773 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hll27\" (UniqueName: \"kubernetes.io/projected/d186292e-15d4-4944-b6ac-7db1fb7b85f5-kube-api-access-hll27\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-cz28q\" (UID: \"d186292e-15d4-4944-b6ac-7db1fb7b85f5\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-cz28q" Dec 01 09:21:49 crc kubenswrapper[4873]: I1201 09:21:49.653868 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d186292e-15d4-4944-b6ac-7db1fb7b85f5-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-cz28q\" (UID: \"d186292e-15d4-4944-b6ac-7db1fb7b85f5\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-cz28q" Dec 01 09:21:49 crc kubenswrapper[4873]: I1201 09:21:49.653869 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d186292e-15d4-4944-b6ac-7db1fb7b85f5-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-cz28q\" (UID: \"d186292e-15d4-4944-b6ac-7db1fb7b85f5\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-cz28q" Dec 01 09:21:49 crc kubenswrapper[4873]: I1201 09:21:49.655782 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d186292e-15d4-4944-b6ac-7db1fb7b85f5-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-cz28q\" (UID: \"d186292e-15d4-4944-b6ac-7db1fb7b85f5\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-cz28q" Dec 01 09:21:49 crc kubenswrapper[4873]: I1201 09:21:49.667764 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hll27\" (UniqueName: \"kubernetes.io/projected/d186292e-15d4-4944-b6ac-7db1fb7b85f5-kube-api-access-hll27\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-cz28q\" (UID: \"d186292e-15d4-4944-b6ac-7db1fb7b85f5\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-cz28q" Dec 01 09:21:49 crc kubenswrapper[4873]: I1201 09:21:49.821854 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-cz28q" Dec 01 09:21:50 crc kubenswrapper[4873]: I1201 09:21:50.374755 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-cz28q"] Dec 01 09:21:51 crc kubenswrapper[4873]: I1201 09:21:51.378658 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-cz28q" event={"ID":"d186292e-15d4-4944-b6ac-7db1fb7b85f5","Type":"ContainerStarted","Data":"66e27a31271e6ab3de66bd1571e84502c738ea958eb6ac4147e4b93454365f32"} Dec 01 09:21:51 crc kubenswrapper[4873]: I1201 09:21:51.379547 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-cz28q" event={"ID":"d186292e-15d4-4944-b6ac-7db1fb7b85f5","Type":"ContainerStarted","Data":"85de498d27a36d3129f38809f6fbd032f1625f91278ae6e035a1b1dbdfe32cb8"} Dec 01 09:21:51 crc kubenswrapper[4873]: I1201 09:21:51.413560 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-cz28q" podStartSLOduration=1.8837861930000002 podStartE2EDuration="2.413535649s" podCreationTimestamp="2025-12-01 09:21:49 +0000 UTC" firstStartedPulling="2025-12-01 09:21:50.379439325 +0000 UTC m=+2486.281547864" lastFinishedPulling="2025-12-01 09:21:50.909188781 +0000 UTC m=+2486.811297320" observedRunningTime="2025-12-01 09:21:51.400491887 +0000 UTC m=+2487.302600426" watchObservedRunningTime="2025-12-01 09:21:51.413535649 +0000 UTC m=+2487.315644188" Dec 01 09:21:55 crc kubenswrapper[4873]: I1201 09:21:55.433435 4873 scope.go:117] "RemoveContainer" containerID="1a1db8f8327e8cf6dbb94b57ea898b250339bb198ae1ba595cb1d79805fa0db1" Dec 01 09:21:55 crc kubenswrapper[4873]: E1201 09:21:55.434730 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:21:57 crc kubenswrapper[4873]: I1201 09:21:57.464598 4873 generic.go:334] "Generic (PLEG): container finished" podID="d186292e-15d4-4944-b6ac-7db1fb7b85f5" containerID="66e27a31271e6ab3de66bd1571e84502c738ea958eb6ac4147e4b93454365f32" exitCode=0 Dec 01 09:21:57 crc kubenswrapper[4873]: I1201 09:21:57.464702 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-cz28q" event={"ID":"d186292e-15d4-4944-b6ac-7db1fb7b85f5","Type":"ContainerDied","Data":"66e27a31271e6ab3de66bd1571e84502c738ea958eb6ac4147e4b93454365f32"} Dec 01 09:21:59 crc kubenswrapper[4873]: I1201 09:21:59.012119 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-cz28q" Dec 01 09:21:59 crc kubenswrapper[4873]: I1201 09:21:59.128432 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d186292e-15d4-4944-b6ac-7db1fb7b85f5-ssh-key\") pod \"d186292e-15d4-4944-b6ac-7db1fb7b85f5\" (UID: \"d186292e-15d4-4944-b6ac-7db1fb7b85f5\") " Dec 01 09:21:59 crc kubenswrapper[4873]: I1201 09:21:59.128506 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d186292e-15d4-4944-b6ac-7db1fb7b85f5-ceph\") pod \"d186292e-15d4-4944-b6ac-7db1fb7b85f5\" (UID: \"d186292e-15d4-4944-b6ac-7db1fb7b85f5\") " Dec 01 09:21:59 crc kubenswrapper[4873]: I1201 09:21:59.128593 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d186292e-15d4-4944-b6ac-7db1fb7b85f5-inventory\") pod \"d186292e-15d4-4944-b6ac-7db1fb7b85f5\" (UID: \"d186292e-15d4-4944-b6ac-7db1fb7b85f5\") " Dec 01 09:21:59 crc kubenswrapper[4873]: I1201 09:21:59.129723 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hll27\" (UniqueName: \"kubernetes.io/projected/d186292e-15d4-4944-b6ac-7db1fb7b85f5-kube-api-access-hll27\") pod \"d186292e-15d4-4944-b6ac-7db1fb7b85f5\" (UID: \"d186292e-15d4-4944-b6ac-7db1fb7b85f5\") " Dec 01 09:21:59 crc kubenswrapper[4873]: I1201 09:21:59.140778 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d186292e-15d4-4944-b6ac-7db1fb7b85f5-kube-api-access-hll27" (OuterVolumeSpecName: "kube-api-access-hll27") pod "d186292e-15d4-4944-b6ac-7db1fb7b85f5" (UID: "d186292e-15d4-4944-b6ac-7db1fb7b85f5"). InnerVolumeSpecName "kube-api-access-hll27". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:21:59 crc kubenswrapper[4873]: I1201 09:21:59.141227 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d186292e-15d4-4944-b6ac-7db1fb7b85f5-ceph" (OuterVolumeSpecName: "ceph") pod "d186292e-15d4-4944-b6ac-7db1fb7b85f5" (UID: "d186292e-15d4-4944-b6ac-7db1fb7b85f5"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:21:59 crc kubenswrapper[4873]: I1201 09:21:59.176786 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d186292e-15d4-4944-b6ac-7db1fb7b85f5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d186292e-15d4-4944-b6ac-7db1fb7b85f5" (UID: "d186292e-15d4-4944-b6ac-7db1fb7b85f5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:21:59 crc kubenswrapper[4873]: I1201 09:21:59.177283 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d186292e-15d4-4944-b6ac-7db1fb7b85f5-inventory" (OuterVolumeSpecName: "inventory") pod "d186292e-15d4-4944-b6ac-7db1fb7b85f5" (UID: "d186292e-15d4-4944-b6ac-7db1fb7b85f5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:21:59 crc kubenswrapper[4873]: I1201 09:21:59.233427 4873 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d186292e-15d4-4944-b6ac-7db1fb7b85f5-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 09:21:59 crc kubenswrapper[4873]: I1201 09:21:59.233473 4873 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d186292e-15d4-4944-b6ac-7db1fb7b85f5-ceph\") on node \"crc\" DevicePath \"\"" Dec 01 09:21:59 crc kubenswrapper[4873]: I1201 09:21:59.233487 4873 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d186292e-15d4-4944-b6ac-7db1fb7b85f5-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 09:21:59 crc kubenswrapper[4873]: I1201 09:21:59.233502 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hll27\" (UniqueName: \"kubernetes.io/projected/d186292e-15d4-4944-b6ac-7db1fb7b85f5-kube-api-access-hll27\") on node \"crc\" DevicePath \"\"" Dec 01 09:21:59 crc kubenswrapper[4873]: I1201 09:21:59.486664 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-cz28q" event={"ID":"d186292e-15d4-4944-b6ac-7db1fb7b85f5","Type":"ContainerDied","Data":"85de498d27a36d3129f38809f6fbd032f1625f91278ae6e035a1b1dbdfe32cb8"} Dec 01 09:21:59 crc kubenswrapper[4873]: I1201 09:21:59.486727 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85de498d27a36d3129f38809f6fbd032f1625f91278ae6e035a1b1dbdfe32cb8" Dec 01 09:21:59 crc kubenswrapper[4873]: I1201 09:21:59.486739 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-cz28q" Dec 01 09:21:59 crc kubenswrapper[4873]: I1201 09:21:59.594973 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-q7cl2"] Dec 01 09:21:59 crc kubenswrapper[4873]: E1201 09:21:59.595663 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d186292e-15d4-4944-b6ac-7db1fb7b85f5" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Dec 01 09:21:59 crc kubenswrapper[4873]: I1201 09:21:59.595693 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="d186292e-15d4-4944-b6ac-7db1fb7b85f5" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Dec 01 09:21:59 crc kubenswrapper[4873]: I1201 09:21:59.596145 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="d186292e-15d4-4944-b6ac-7db1fb7b85f5" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Dec 01 09:21:59 crc kubenswrapper[4873]: I1201 09:21:59.597130 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q7cl2" Dec 01 09:21:59 crc kubenswrapper[4873]: I1201 09:21:59.601403 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 01 09:21:59 crc kubenswrapper[4873]: I1201 09:21:59.601833 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9ghqw" Dec 01 09:21:59 crc kubenswrapper[4873]: I1201 09:21:59.602106 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 01 09:21:59 crc kubenswrapper[4873]: I1201 09:21:59.602258 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 09:21:59 crc kubenswrapper[4873]: I1201 09:21:59.602349 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 09:21:59 crc kubenswrapper[4873]: I1201 09:21:59.602439 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 09:21:59 crc kubenswrapper[4873]: I1201 09:21:59.605044 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-q7cl2"] Dec 01 09:21:59 crc kubenswrapper[4873]: I1201 09:21:59.753954 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/0b014b12-82de-4f51-84bc-c8d41241f672-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q7cl2\" (UID: \"0b014b12-82de-4f51-84bc-c8d41241f672\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q7cl2" Dec 01 09:21:59 crc kubenswrapper[4873]: I1201 09:21:59.754194 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0b014b12-82de-4f51-84bc-c8d41241f672-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q7cl2\" (UID: \"0b014b12-82de-4f51-84bc-c8d41241f672\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q7cl2" Dec 01 09:21:59 crc kubenswrapper[4873]: I1201 09:21:59.754261 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fv26h\" (UniqueName: \"kubernetes.io/projected/0b014b12-82de-4f51-84bc-c8d41241f672-kube-api-access-fv26h\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q7cl2\" (UID: \"0b014b12-82de-4f51-84bc-c8d41241f672\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q7cl2" Dec 01 09:21:59 crc kubenswrapper[4873]: I1201 09:21:59.754370 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b014b12-82de-4f51-84bc-c8d41241f672-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q7cl2\" (UID: \"0b014b12-82de-4f51-84bc-c8d41241f672\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q7cl2" Dec 01 09:21:59 crc kubenswrapper[4873]: I1201 09:21:59.754547 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0b014b12-82de-4f51-84bc-c8d41241f672-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q7cl2\" (UID: \"0b014b12-82de-4f51-84bc-c8d41241f672\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q7cl2" Dec 01 09:21:59 crc kubenswrapper[4873]: I1201 09:21:59.754647 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0b014b12-82de-4f51-84bc-c8d41241f672-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q7cl2\" (UID: \"0b014b12-82de-4f51-84bc-c8d41241f672\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q7cl2" Dec 01 09:21:59 crc kubenswrapper[4873]: I1201 09:21:59.857069 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0b014b12-82de-4f51-84bc-c8d41241f672-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q7cl2\" (UID: \"0b014b12-82de-4f51-84bc-c8d41241f672\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q7cl2" Dec 01 09:21:59 crc kubenswrapper[4873]: I1201 09:21:59.857311 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/0b014b12-82de-4f51-84bc-c8d41241f672-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q7cl2\" (UID: \"0b014b12-82de-4f51-84bc-c8d41241f672\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q7cl2" Dec 01 09:21:59 crc kubenswrapper[4873]: I1201 09:21:59.857354 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0b014b12-82de-4f51-84bc-c8d41241f672-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q7cl2\" (UID: \"0b014b12-82de-4f51-84bc-c8d41241f672\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q7cl2" Dec 01 09:21:59 crc kubenswrapper[4873]: I1201 09:21:59.857395 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fv26h\" (UniqueName: \"kubernetes.io/projected/0b014b12-82de-4f51-84bc-c8d41241f672-kube-api-access-fv26h\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q7cl2\" (UID: \"0b014b12-82de-4f51-84bc-c8d41241f672\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q7cl2" Dec 01 09:21:59 crc kubenswrapper[4873]: I1201 09:21:59.857431 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b014b12-82de-4f51-84bc-c8d41241f672-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q7cl2\" (UID: \"0b014b12-82de-4f51-84bc-c8d41241f672\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q7cl2" Dec 01 09:21:59 crc kubenswrapper[4873]: I1201 09:21:59.857518 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0b014b12-82de-4f51-84bc-c8d41241f672-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q7cl2\" (UID: \"0b014b12-82de-4f51-84bc-c8d41241f672\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q7cl2" Dec 01 09:21:59 crc kubenswrapper[4873]: I1201 09:21:59.858698 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/0b014b12-82de-4f51-84bc-c8d41241f672-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q7cl2\" (UID: \"0b014b12-82de-4f51-84bc-c8d41241f672\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q7cl2" Dec 01 09:21:59 crc kubenswrapper[4873]: I1201 09:21:59.862324 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0b014b12-82de-4f51-84bc-c8d41241f672-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q7cl2\" (UID: \"0b014b12-82de-4f51-84bc-c8d41241f672\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q7cl2" Dec 01 09:21:59 crc kubenswrapper[4873]: I1201 09:21:59.863108 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b014b12-82de-4f51-84bc-c8d41241f672-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q7cl2\" (UID: \"0b014b12-82de-4f51-84bc-c8d41241f672\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q7cl2" Dec 01 09:21:59 crc kubenswrapper[4873]: I1201 09:21:59.863502 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0b014b12-82de-4f51-84bc-c8d41241f672-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q7cl2\" (UID: \"0b014b12-82de-4f51-84bc-c8d41241f672\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q7cl2" Dec 01 09:21:59 crc kubenswrapper[4873]: I1201 09:21:59.863527 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0b014b12-82de-4f51-84bc-c8d41241f672-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q7cl2\" (UID: \"0b014b12-82de-4f51-84bc-c8d41241f672\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q7cl2" Dec 01 09:21:59 crc kubenswrapper[4873]: I1201 09:21:59.881439 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fv26h\" (UniqueName: \"kubernetes.io/projected/0b014b12-82de-4f51-84bc-c8d41241f672-kube-api-access-fv26h\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q7cl2\" (UID: \"0b014b12-82de-4f51-84bc-c8d41241f672\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q7cl2" Dec 01 09:21:59 crc kubenswrapper[4873]: I1201 09:21:59.938778 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q7cl2" Dec 01 09:22:00 crc kubenswrapper[4873]: I1201 09:22:00.666959 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-q7cl2"] Dec 01 09:22:01 crc kubenswrapper[4873]: I1201 09:22:01.510350 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q7cl2" event={"ID":"0b014b12-82de-4f51-84bc-c8d41241f672","Type":"ContainerStarted","Data":"ea57ce8bb360fee4d34f426596f67c7bef03dbebc54112e65de13b0fe4ba87df"} Dec 01 09:22:02 crc kubenswrapper[4873]: I1201 09:22:02.521458 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q7cl2" event={"ID":"0b014b12-82de-4f51-84bc-c8d41241f672","Type":"ContainerStarted","Data":"e05a704f6f702472ecc1b1c1ccdce08434f18820270f4688d60908ebe7442b3d"} Dec 01 09:22:02 crc kubenswrapper[4873]: I1201 09:22:02.545728 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q7cl2" podStartSLOduration=2.33518298 podStartE2EDuration="3.545705503s" podCreationTimestamp="2025-12-01 09:21:59 +0000 UTC" firstStartedPulling="2025-12-01 09:22:00.673597418 +0000 UTC m=+2496.575705957" lastFinishedPulling="2025-12-01 09:22:01.884119941 +0000 UTC m=+2497.786228480" observedRunningTime="2025-12-01 09:22:02.539731616 +0000 UTC m=+2498.441840155" watchObservedRunningTime="2025-12-01 09:22:02.545705503 +0000 UTC m=+2498.447814032" Dec 01 09:22:07 crc kubenswrapper[4873]: I1201 09:22:07.430774 4873 scope.go:117] "RemoveContainer" containerID="1a1db8f8327e8cf6dbb94b57ea898b250339bb198ae1ba595cb1d79805fa0db1" Dec 01 09:22:08 crc kubenswrapper[4873]: I1201 09:22:08.578692 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" event={"ID":"fef7b114-0e07-402d-a37b-315c36011f4b","Type":"ContainerStarted","Data":"5086f7da0c61e3334bc2085e8599b53512aec9e3eb10321166ea8d6e4ebb4c9d"} Dec 01 09:23:16 crc kubenswrapper[4873]: I1201 09:23:16.244304 4873 generic.go:334] "Generic (PLEG): container finished" podID="0b014b12-82de-4f51-84bc-c8d41241f672" containerID="e05a704f6f702472ecc1b1c1ccdce08434f18820270f4688d60908ebe7442b3d" exitCode=0 Dec 01 09:23:16 crc kubenswrapper[4873]: I1201 09:23:16.244383 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q7cl2" event={"ID":"0b014b12-82de-4f51-84bc-c8d41241f672","Type":"ContainerDied","Data":"e05a704f6f702472ecc1b1c1ccdce08434f18820270f4688d60908ebe7442b3d"} Dec 01 09:23:17 crc kubenswrapper[4873]: I1201 09:23:17.762184 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q7cl2" Dec 01 09:23:17 crc kubenswrapper[4873]: I1201 09:23:17.879142 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0b014b12-82de-4f51-84bc-c8d41241f672-ssh-key\") pod \"0b014b12-82de-4f51-84bc-c8d41241f672\" (UID: \"0b014b12-82de-4f51-84bc-c8d41241f672\") " Dec 01 09:23:17 crc kubenswrapper[4873]: I1201 09:23:17.879200 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b014b12-82de-4f51-84bc-c8d41241f672-ovn-combined-ca-bundle\") pod \"0b014b12-82de-4f51-84bc-c8d41241f672\" (UID: \"0b014b12-82de-4f51-84bc-c8d41241f672\") " Dec 01 09:23:17 crc kubenswrapper[4873]: I1201 09:23:17.879246 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0b014b12-82de-4f51-84bc-c8d41241f672-inventory\") pod \"0b014b12-82de-4f51-84bc-c8d41241f672\" (UID: \"0b014b12-82de-4f51-84bc-c8d41241f672\") " Dec 01 09:23:17 crc kubenswrapper[4873]: I1201 09:23:17.879298 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0b014b12-82de-4f51-84bc-c8d41241f672-ceph\") pod \"0b014b12-82de-4f51-84bc-c8d41241f672\" (UID: \"0b014b12-82de-4f51-84bc-c8d41241f672\") " Dec 01 09:23:17 crc kubenswrapper[4873]: I1201 09:23:17.879406 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/0b014b12-82de-4f51-84bc-c8d41241f672-ovncontroller-config-0\") pod \"0b014b12-82de-4f51-84bc-c8d41241f672\" (UID: \"0b014b12-82de-4f51-84bc-c8d41241f672\") " Dec 01 09:23:17 crc kubenswrapper[4873]: I1201 09:23:17.879577 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fv26h\" (UniqueName: \"kubernetes.io/projected/0b014b12-82de-4f51-84bc-c8d41241f672-kube-api-access-fv26h\") pod \"0b014b12-82de-4f51-84bc-c8d41241f672\" (UID: \"0b014b12-82de-4f51-84bc-c8d41241f672\") " Dec 01 09:23:17 crc kubenswrapper[4873]: I1201 09:23:17.885376 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b014b12-82de-4f51-84bc-c8d41241f672-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "0b014b12-82de-4f51-84bc-c8d41241f672" (UID: "0b014b12-82de-4f51-84bc-c8d41241f672"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:23:17 crc kubenswrapper[4873]: I1201 09:23:17.885745 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b014b12-82de-4f51-84bc-c8d41241f672-ceph" (OuterVolumeSpecName: "ceph") pod "0b014b12-82de-4f51-84bc-c8d41241f672" (UID: "0b014b12-82de-4f51-84bc-c8d41241f672"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:23:17 crc kubenswrapper[4873]: I1201 09:23:17.886063 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b014b12-82de-4f51-84bc-c8d41241f672-kube-api-access-fv26h" (OuterVolumeSpecName: "kube-api-access-fv26h") pod "0b014b12-82de-4f51-84bc-c8d41241f672" (UID: "0b014b12-82de-4f51-84bc-c8d41241f672"). InnerVolumeSpecName "kube-api-access-fv26h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:23:17 crc kubenswrapper[4873]: I1201 09:23:17.907768 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b014b12-82de-4f51-84bc-c8d41241f672-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0b014b12-82de-4f51-84bc-c8d41241f672" (UID: "0b014b12-82de-4f51-84bc-c8d41241f672"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:23:17 crc kubenswrapper[4873]: I1201 09:23:17.911736 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b014b12-82de-4f51-84bc-c8d41241f672-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "0b014b12-82de-4f51-84bc-c8d41241f672" (UID: "0b014b12-82de-4f51-84bc-c8d41241f672"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:23:17 crc kubenswrapper[4873]: I1201 09:23:17.913919 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b014b12-82de-4f51-84bc-c8d41241f672-inventory" (OuterVolumeSpecName: "inventory") pod "0b014b12-82de-4f51-84bc-c8d41241f672" (UID: "0b014b12-82de-4f51-84bc-c8d41241f672"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:23:17 crc kubenswrapper[4873]: I1201 09:23:17.983951 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fv26h\" (UniqueName: \"kubernetes.io/projected/0b014b12-82de-4f51-84bc-c8d41241f672-kube-api-access-fv26h\") on node \"crc\" DevicePath \"\"" Dec 01 09:23:17 crc kubenswrapper[4873]: I1201 09:23:17.984608 4873 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0b014b12-82de-4f51-84bc-c8d41241f672-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 09:23:17 crc kubenswrapper[4873]: I1201 09:23:17.984719 4873 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b014b12-82de-4f51-84bc-c8d41241f672-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:23:17 crc kubenswrapper[4873]: I1201 09:23:17.984810 4873 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0b014b12-82de-4f51-84bc-c8d41241f672-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 09:23:17 crc kubenswrapper[4873]: I1201 09:23:17.984917 4873 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0b014b12-82de-4f51-84bc-c8d41241f672-ceph\") on node \"crc\" DevicePath \"\"" Dec 01 09:23:17 crc kubenswrapper[4873]: I1201 09:23:17.985007 4873 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/0b014b12-82de-4f51-84bc-c8d41241f672-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 01 09:23:18 crc kubenswrapper[4873]: I1201 09:23:18.267289 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q7cl2" event={"ID":"0b014b12-82de-4f51-84bc-c8d41241f672","Type":"ContainerDied","Data":"ea57ce8bb360fee4d34f426596f67c7bef03dbebc54112e65de13b0fe4ba87df"} Dec 01 09:23:18 crc kubenswrapper[4873]: I1201 09:23:18.267345 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea57ce8bb360fee4d34f426596f67c7bef03dbebc54112e65de13b0fe4ba87df" Dec 01 09:23:18 crc kubenswrapper[4873]: I1201 09:23:18.267442 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q7cl2" Dec 01 09:23:18 crc kubenswrapper[4873]: I1201 09:23:18.424456 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b"] Dec 01 09:23:18 crc kubenswrapper[4873]: E1201 09:23:18.426216 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b014b12-82de-4f51-84bc-c8d41241f672" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 01 09:23:18 crc kubenswrapper[4873]: I1201 09:23:18.426298 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b014b12-82de-4f51-84bc-c8d41241f672" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 01 09:23:18 crc kubenswrapper[4873]: I1201 09:23:18.427025 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b014b12-82de-4f51-84bc-c8d41241f672" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 01 09:23:18 crc kubenswrapper[4873]: I1201 09:23:18.428389 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b" Dec 01 09:23:18 crc kubenswrapper[4873]: I1201 09:23:18.437899 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 01 09:23:18 crc kubenswrapper[4873]: I1201 09:23:18.438492 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 09:23:18 crc kubenswrapper[4873]: I1201 09:23:18.438714 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 09:23:18 crc kubenswrapper[4873]: I1201 09:23:18.440629 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9ghqw" Dec 01 09:23:18 crc kubenswrapper[4873]: I1201 09:23:18.441449 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 01 09:23:18 crc kubenswrapper[4873]: I1201 09:23:18.441858 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 01 09:23:18 crc kubenswrapper[4873]: I1201 09:23:18.446347 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 09:23:18 crc kubenswrapper[4873]: I1201 09:23:18.476421 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b"] Dec 01 09:23:18 crc kubenswrapper[4873]: I1201 09:23:18.597279 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ed4f574a-846e-41be-88ce-56d89a392a87-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b\" (UID: \"ed4f574a-846e-41be-88ce-56d89a392a87\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b" Dec 01 09:23:18 crc kubenswrapper[4873]: I1201 09:23:18.597357 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ed4f574a-846e-41be-88ce-56d89a392a87-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b\" (UID: \"ed4f574a-846e-41be-88ce-56d89a392a87\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b" Dec 01 09:23:18 crc kubenswrapper[4873]: I1201 09:23:18.597400 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed4f574a-846e-41be-88ce-56d89a392a87-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b\" (UID: \"ed4f574a-846e-41be-88ce-56d89a392a87\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b" Dec 01 09:23:18 crc kubenswrapper[4873]: I1201 09:23:18.597441 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgjdj\" (UniqueName: \"kubernetes.io/projected/ed4f574a-846e-41be-88ce-56d89a392a87-kube-api-access-cgjdj\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b\" (UID: \"ed4f574a-846e-41be-88ce-56d89a392a87\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b" Dec 01 09:23:18 crc kubenswrapper[4873]: I1201 09:23:18.597492 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ed4f574a-846e-41be-88ce-56d89a392a87-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b\" (UID: \"ed4f574a-846e-41be-88ce-56d89a392a87\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b" Dec 01 09:23:18 crc kubenswrapper[4873]: I1201 09:23:18.597645 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ed4f574a-846e-41be-88ce-56d89a392a87-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b\" (UID: \"ed4f574a-846e-41be-88ce-56d89a392a87\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b" Dec 01 09:23:18 crc kubenswrapper[4873]: I1201 09:23:18.597786 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ed4f574a-846e-41be-88ce-56d89a392a87-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b\" (UID: \"ed4f574a-846e-41be-88ce-56d89a392a87\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b" Dec 01 09:23:18 crc kubenswrapper[4873]: I1201 09:23:18.700309 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed4f574a-846e-41be-88ce-56d89a392a87-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b\" (UID: \"ed4f574a-846e-41be-88ce-56d89a392a87\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b" Dec 01 09:23:18 crc kubenswrapper[4873]: I1201 09:23:18.700376 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ed4f574a-846e-41be-88ce-56d89a392a87-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b\" (UID: \"ed4f574a-846e-41be-88ce-56d89a392a87\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b" Dec 01 09:23:18 crc kubenswrapper[4873]: I1201 09:23:18.700423 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgjdj\" (UniqueName: \"kubernetes.io/projected/ed4f574a-846e-41be-88ce-56d89a392a87-kube-api-access-cgjdj\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b\" (UID: \"ed4f574a-846e-41be-88ce-56d89a392a87\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b" Dec 01 09:23:18 crc kubenswrapper[4873]: I1201 09:23:18.700486 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ed4f574a-846e-41be-88ce-56d89a392a87-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b\" (UID: \"ed4f574a-846e-41be-88ce-56d89a392a87\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b" Dec 01 09:23:18 crc kubenswrapper[4873]: I1201 09:23:18.700554 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ed4f574a-846e-41be-88ce-56d89a392a87-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b\" (UID: \"ed4f574a-846e-41be-88ce-56d89a392a87\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b" Dec 01 09:23:18 crc kubenswrapper[4873]: I1201 09:23:18.700766 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ed4f574a-846e-41be-88ce-56d89a392a87-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b\" (UID: \"ed4f574a-846e-41be-88ce-56d89a392a87\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b" Dec 01 09:23:18 crc kubenswrapper[4873]: I1201 09:23:18.700915 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ed4f574a-846e-41be-88ce-56d89a392a87-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b\" (UID: \"ed4f574a-846e-41be-88ce-56d89a392a87\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b" Dec 01 09:23:18 crc kubenswrapper[4873]: I1201 09:23:18.706300 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ed4f574a-846e-41be-88ce-56d89a392a87-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b\" (UID: \"ed4f574a-846e-41be-88ce-56d89a392a87\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b" Dec 01 09:23:18 crc kubenswrapper[4873]: I1201 09:23:18.706487 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ed4f574a-846e-41be-88ce-56d89a392a87-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b\" (UID: \"ed4f574a-846e-41be-88ce-56d89a392a87\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b" Dec 01 09:23:18 crc kubenswrapper[4873]: I1201 09:23:18.706904 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ed4f574a-846e-41be-88ce-56d89a392a87-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b\" (UID: \"ed4f574a-846e-41be-88ce-56d89a392a87\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b" Dec 01 09:23:18 crc kubenswrapper[4873]: I1201 09:23:18.707394 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed4f574a-846e-41be-88ce-56d89a392a87-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b\" (UID: \"ed4f574a-846e-41be-88ce-56d89a392a87\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b" Dec 01 09:23:18 crc kubenswrapper[4873]: I1201 09:23:18.708405 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ed4f574a-846e-41be-88ce-56d89a392a87-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b\" (UID: \"ed4f574a-846e-41be-88ce-56d89a392a87\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b" Dec 01 09:23:18 crc kubenswrapper[4873]: I1201 09:23:18.712804 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ed4f574a-846e-41be-88ce-56d89a392a87-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b\" (UID: \"ed4f574a-846e-41be-88ce-56d89a392a87\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b" Dec 01 09:23:18 crc kubenswrapper[4873]: I1201 09:23:18.719318 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgjdj\" (UniqueName: \"kubernetes.io/projected/ed4f574a-846e-41be-88ce-56d89a392a87-kube-api-access-cgjdj\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b\" (UID: \"ed4f574a-846e-41be-88ce-56d89a392a87\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b" Dec 01 09:23:18 crc kubenswrapper[4873]: I1201 09:23:18.761375 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b" Dec 01 09:23:19 crc kubenswrapper[4873]: I1201 09:23:19.295742 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b"] Dec 01 09:23:20 crc kubenswrapper[4873]: I1201 09:23:20.289265 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b" event={"ID":"ed4f574a-846e-41be-88ce-56d89a392a87","Type":"ContainerStarted","Data":"65127fc1735f3360e0f9c8e2e73ab831835281984852d208d2406e9c594d00ef"} Dec 01 09:23:20 crc kubenswrapper[4873]: I1201 09:23:20.290149 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b" event={"ID":"ed4f574a-846e-41be-88ce-56d89a392a87","Type":"ContainerStarted","Data":"7f20ddf0e300e152114d8d64f52389b07194a3f67ed07e6ce62a137b22a5ba14"} Dec 01 09:23:20 crc kubenswrapper[4873]: I1201 09:23:20.313037 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b" podStartSLOduration=1.773828519 podStartE2EDuration="2.313008658s" podCreationTimestamp="2025-12-01 09:23:18 +0000 UTC" firstStartedPulling="2025-12-01 09:23:19.292621032 +0000 UTC m=+2575.194729571" lastFinishedPulling="2025-12-01 09:23:19.831801171 +0000 UTC m=+2575.733909710" observedRunningTime="2025-12-01 09:23:20.312402413 +0000 UTC m=+2576.214510972" watchObservedRunningTime="2025-12-01 09:23:20.313008658 +0000 UTC m=+2576.215117197" Dec 01 09:24:19 crc kubenswrapper[4873]: I1201 09:24:19.916982 4873 generic.go:334] "Generic (PLEG): container finished" podID="ed4f574a-846e-41be-88ce-56d89a392a87" containerID="65127fc1735f3360e0f9c8e2e73ab831835281984852d208d2406e9c594d00ef" exitCode=0 Dec 01 09:24:19 crc kubenswrapper[4873]: I1201 09:24:19.917051 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b" event={"ID":"ed4f574a-846e-41be-88ce-56d89a392a87","Type":"ContainerDied","Data":"65127fc1735f3360e0f9c8e2e73ab831835281984852d208d2406e9c594d00ef"} Dec 01 09:24:21 crc kubenswrapper[4873]: I1201 09:24:21.365403 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b" Dec 01 09:24:21 crc kubenswrapper[4873]: I1201 09:24:21.498985 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cgjdj\" (UniqueName: \"kubernetes.io/projected/ed4f574a-846e-41be-88ce-56d89a392a87-kube-api-access-cgjdj\") pod \"ed4f574a-846e-41be-88ce-56d89a392a87\" (UID: \"ed4f574a-846e-41be-88ce-56d89a392a87\") " Dec 01 09:24:21 crc kubenswrapper[4873]: I1201 09:24:21.499163 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed4f574a-846e-41be-88ce-56d89a392a87-neutron-metadata-combined-ca-bundle\") pod \"ed4f574a-846e-41be-88ce-56d89a392a87\" (UID: \"ed4f574a-846e-41be-88ce-56d89a392a87\") " Dec 01 09:24:21 crc kubenswrapper[4873]: I1201 09:24:21.499269 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ed4f574a-846e-41be-88ce-56d89a392a87-ceph\") pod \"ed4f574a-846e-41be-88ce-56d89a392a87\" (UID: \"ed4f574a-846e-41be-88ce-56d89a392a87\") " Dec 01 09:24:21 crc kubenswrapper[4873]: I1201 09:24:21.499333 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ed4f574a-846e-41be-88ce-56d89a392a87-neutron-ovn-metadata-agent-neutron-config-0\") pod \"ed4f574a-846e-41be-88ce-56d89a392a87\" (UID: \"ed4f574a-846e-41be-88ce-56d89a392a87\") " Dec 01 09:24:21 crc kubenswrapper[4873]: I1201 09:24:21.499385 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ed4f574a-846e-41be-88ce-56d89a392a87-nova-metadata-neutron-config-0\") pod \"ed4f574a-846e-41be-88ce-56d89a392a87\" (UID: \"ed4f574a-846e-41be-88ce-56d89a392a87\") " Dec 01 09:24:21 crc kubenswrapper[4873]: I1201 09:24:21.499416 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ed4f574a-846e-41be-88ce-56d89a392a87-ssh-key\") pod \"ed4f574a-846e-41be-88ce-56d89a392a87\" (UID: \"ed4f574a-846e-41be-88ce-56d89a392a87\") " Dec 01 09:24:21 crc kubenswrapper[4873]: I1201 09:24:21.499447 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ed4f574a-846e-41be-88ce-56d89a392a87-inventory\") pod \"ed4f574a-846e-41be-88ce-56d89a392a87\" (UID: \"ed4f574a-846e-41be-88ce-56d89a392a87\") " Dec 01 09:24:21 crc kubenswrapper[4873]: I1201 09:24:21.505499 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed4f574a-846e-41be-88ce-56d89a392a87-ceph" (OuterVolumeSpecName: "ceph") pod "ed4f574a-846e-41be-88ce-56d89a392a87" (UID: "ed4f574a-846e-41be-88ce-56d89a392a87"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:24:21 crc kubenswrapper[4873]: I1201 09:24:21.505992 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed4f574a-846e-41be-88ce-56d89a392a87-kube-api-access-cgjdj" (OuterVolumeSpecName: "kube-api-access-cgjdj") pod "ed4f574a-846e-41be-88ce-56d89a392a87" (UID: "ed4f574a-846e-41be-88ce-56d89a392a87"). InnerVolumeSpecName "kube-api-access-cgjdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:24:21 crc kubenswrapper[4873]: I1201 09:24:21.507349 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed4f574a-846e-41be-88ce-56d89a392a87-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "ed4f574a-846e-41be-88ce-56d89a392a87" (UID: "ed4f574a-846e-41be-88ce-56d89a392a87"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:24:21 crc kubenswrapper[4873]: I1201 09:24:21.530531 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed4f574a-846e-41be-88ce-56d89a392a87-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "ed4f574a-846e-41be-88ce-56d89a392a87" (UID: "ed4f574a-846e-41be-88ce-56d89a392a87"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:24:21 crc kubenswrapper[4873]: I1201 09:24:21.541280 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed4f574a-846e-41be-88ce-56d89a392a87-inventory" (OuterVolumeSpecName: "inventory") pod "ed4f574a-846e-41be-88ce-56d89a392a87" (UID: "ed4f574a-846e-41be-88ce-56d89a392a87"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:24:21 crc kubenswrapper[4873]: I1201 09:24:21.541494 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed4f574a-846e-41be-88ce-56d89a392a87-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "ed4f574a-846e-41be-88ce-56d89a392a87" (UID: "ed4f574a-846e-41be-88ce-56d89a392a87"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:24:21 crc kubenswrapper[4873]: I1201 09:24:21.542222 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed4f574a-846e-41be-88ce-56d89a392a87-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ed4f574a-846e-41be-88ce-56d89a392a87" (UID: "ed4f574a-846e-41be-88ce-56d89a392a87"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:24:21 crc kubenswrapper[4873]: I1201 09:24:21.602881 4873 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ed4f574a-846e-41be-88ce-56d89a392a87-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 09:24:21 crc kubenswrapper[4873]: I1201 09:24:21.602928 4873 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ed4f574a-846e-41be-88ce-56d89a392a87-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 09:24:21 crc kubenswrapper[4873]: I1201 09:24:21.602942 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cgjdj\" (UniqueName: \"kubernetes.io/projected/ed4f574a-846e-41be-88ce-56d89a392a87-kube-api-access-cgjdj\") on node \"crc\" DevicePath \"\"" Dec 01 09:24:21 crc kubenswrapper[4873]: I1201 09:24:21.602968 4873 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed4f574a-846e-41be-88ce-56d89a392a87-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:24:21 crc kubenswrapper[4873]: I1201 09:24:21.602982 4873 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ed4f574a-846e-41be-88ce-56d89a392a87-ceph\") on node \"crc\" DevicePath \"\"" Dec 01 09:24:21 crc kubenswrapper[4873]: I1201 09:24:21.602996 4873 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ed4f574a-846e-41be-88ce-56d89a392a87-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 01 09:24:21 crc kubenswrapper[4873]: I1201 09:24:21.603010 4873 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ed4f574a-846e-41be-88ce-56d89a392a87-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 01 09:24:21 crc kubenswrapper[4873]: I1201 09:24:21.940069 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b" event={"ID":"ed4f574a-846e-41be-88ce-56d89a392a87","Type":"ContainerDied","Data":"7f20ddf0e300e152114d8d64f52389b07194a3f67ed07e6ce62a137b22a5ba14"} Dec 01 09:24:21 crc kubenswrapper[4873]: I1201 09:24:21.940470 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f20ddf0e300e152114d8d64f52389b07194a3f67ed07e6ce62a137b22a5ba14" Dec 01 09:24:21 crc kubenswrapper[4873]: I1201 09:24:21.940135 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b" Dec 01 09:24:22 crc kubenswrapper[4873]: I1201 09:24:22.047828 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp"] Dec 01 09:24:22 crc kubenswrapper[4873]: E1201 09:24:22.048408 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed4f574a-846e-41be-88ce-56d89a392a87" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 01 09:24:22 crc kubenswrapper[4873]: I1201 09:24:22.048434 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed4f574a-846e-41be-88ce-56d89a392a87" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 01 09:24:22 crc kubenswrapper[4873]: I1201 09:24:22.048628 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed4f574a-846e-41be-88ce-56d89a392a87" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 01 09:24:22 crc kubenswrapper[4873]: I1201 09:24:22.049457 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp" Dec 01 09:24:22 crc kubenswrapper[4873]: I1201 09:24:22.053883 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 01 09:24:22 crc kubenswrapper[4873]: I1201 09:24:22.054151 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 01 09:24:22 crc kubenswrapper[4873]: I1201 09:24:22.054319 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 09:24:22 crc kubenswrapper[4873]: I1201 09:24:22.054520 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 09:24:22 crc kubenswrapper[4873]: I1201 09:24:22.054686 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 09:24:22 crc kubenswrapper[4873]: I1201 09:24:22.054840 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9ghqw" Dec 01 09:24:22 crc kubenswrapper[4873]: I1201 09:24:22.080217 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp"] Dec 01 09:24:22 crc kubenswrapper[4873]: I1201 09:24:22.112673 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/328559dd-d6ca-400a-9bb1-955781e2e1ea-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp\" (UID: \"328559dd-d6ca-400a-9bb1-955781e2e1ea\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp" Dec 01 09:24:22 crc kubenswrapper[4873]: I1201 09:24:22.112729 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/328559dd-d6ca-400a-9bb1-955781e2e1ea-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp\" (UID: \"328559dd-d6ca-400a-9bb1-955781e2e1ea\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp" Dec 01 09:24:22 crc kubenswrapper[4873]: I1201 09:24:22.112755 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/328559dd-d6ca-400a-9bb1-955781e2e1ea-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp\" (UID: \"328559dd-d6ca-400a-9bb1-955781e2e1ea\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp" Dec 01 09:24:22 crc kubenswrapper[4873]: I1201 09:24:22.112784 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdcxs\" (UniqueName: \"kubernetes.io/projected/328559dd-d6ca-400a-9bb1-955781e2e1ea-kube-api-access-qdcxs\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp\" (UID: \"328559dd-d6ca-400a-9bb1-955781e2e1ea\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp" Dec 01 09:24:22 crc kubenswrapper[4873]: I1201 09:24:22.112807 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/328559dd-d6ca-400a-9bb1-955781e2e1ea-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp\" (UID: \"328559dd-d6ca-400a-9bb1-955781e2e1ea\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp" Dec 01 09:24:22 crc kubenswrapper[4873]: I1201 09:24:22.113230 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/328559dd-d6ca-400a-9bb1-955781e2e1ea-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp\" (UID: \"328559dd-d6ca-400a-9bb1-955781e2e1ea\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp" Dec 01 09:24:22 crc kubenswrapper[4873]: I1201 09:24:22.214558 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/328559dd-d6ca-400a-9bb1-955781e2e1ea-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp\" (UID: \"328559dd-d6ca-400a-9bb1-955781e2e1ea\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp" Dec 01 09:24:22 crc kubenswrapper[4873]: I1201 09:24:22.214626 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/328559dd-d6ca-400a-9bb1-955781e2e1ea-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp\" (UID: \"328559dd-d6ca-400a-9bb1-955781e2e1ea\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp" Dec 01 09:24:22 crc kubenswrapper[4873]: I1201 09:24:22.214659 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/328559dd-d6ca-400a-9bb1-955781e2e1ea-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp\" (UID: \"328559dd-d6ca-400a-9bb1-955781e2e1ea\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp" Dec 01 09:24:22 crc kubenswrapper[4873]: I1201 09:24:22.214681 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/328559dd-d6ca-400a-9bb1-955781e2e1ea-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp\" (UID: \"328559dd-d6ca-400a-9bb1-955781e2e1ea\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp" Dec 01 09:24:22 crc kubenswrapper[4873]: I1201 09:24:22.214701 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdcxs\" (UniqueName: \"kubernetes.io/projected/328559dd-d6ca-400a-9bb1-955781e2e1ea-kube-api-access-qdcxs\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp\" (UID: \"328559dd-d6ca-400a-9bb1-955781e2e1ea\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp" Dec 01 09:24:22 crc kubenswrapper[4873]: I1201 09:24:22.214719 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/328559dd-d6ca-400a-9bb1-955781e2e1ea-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp\" (UID: \"328559dd-d6ca-400a-9bb1-955781e2e1ea\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp" Dec 01 09:24:22 crc kubenswrapper[4873]: I1201 09:24:22.220176 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/328559dd-d6ca-400a-9bb1-955781e2e1ea-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp\" (UID: \"328559dd-d6ca-400a-9bb1-955781e2e1ea\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp" Dec 01 09:24:22 crc kubenswrapper[4873]: I1201 09:24:22.220208 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/328559dd-d6ca-400a-9bb1-955781e2e1ea-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp\" (UID: \"328559dd-d6ca-400a-9bb1-955781e2e1ea\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp" Dec 01 09:24:22 crc kubenswrapper[4873]: I1201 09:24:22.221057 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/328559dd-d6ca-400a-9bb1-955781e2e1ea-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp\" (UID: \"328559dd-d6ca-400a-9bb1-955781e2e1ea\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp" Dec 01 09:24:22 crc kubenswrapper[4873]: I1201 09:24:22.221401 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/328559dd-d6ca-400a-9bb1-955781e2e1ea-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp\" (UID: \"328559dd-d6ca-400a-9bb1-955781e2e1ea\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp" Dec 01 09:24:22 crc kubenswrapper[4873]: I1201 09:24:22.222910 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/328559dd-d6ca-400a-9bb1-955781e2e1ea-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp\" (UID: \"328559dd-d6ca-400a-9bb1-955781e2e1ea\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp" Dec 01 09:24:22 crc kubenswrapper[4873]: I1201 09:24:22.236171 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdcxs\" (UniqueName: \"kubernetes.io/projected/328559dd-d6ca-400a-9bb1-955781e2e1ea-kube-api-access-qdcxs\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp\" (UID: \"328559dd-d6ca-400a-9bb1-955781e2e1ea\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp" Dec 01 09:24:22 crc kubenswrapper[4873]: I1201 09:24:22.376740 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp" Dec 01 09:24:22 crc kubenswrapper[4873]: I1201 09:24:22.966303 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp"] Dec 01 09:24:23 crc kubenswrapper[4873]: I1201 09:24:23.962587 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp" event={"ID":"328559dd-d6ca-400a-9bb1-955781e2e1ea","Type":"ContainerStarted","Data":"a88250fe25d343a65ff9649e1baedf912e4e13be7f730b6d1a0c8992e987e074"} Dec 01 09:24:24 crc kubenswrapper[4873]: I1201 09:24:24.975565 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp" event={"ID":"328559dd-d6ca-400a-9bb1-955781e2e1ea","Type":"ContainerStarted","Data":"c834e261924bb4cd420b71da1622d43fbc94e7692cbe694dc457544c4c9b46bb"} Dec 01 09:24:24 crc kubenswrapper[4873]: I1201 09:24:24.997780 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp" podStartSLOduration=2.102998601 podStartE2EDuration="2.997751547s" podCreationTimestamp="2025-12-01 09:24:22 +0000 UTC" firstStartedPulling="2025-12-01 09:24:22.975183607 +0000 UTC m=+2638.877292146" lastFinishedPulling="2025-12-01 09:24:23.869936553 +0000 UTC m=+2639.772045092" observedRunningTime="2025-12-01 09:24:24.996810844 +0000 UTC m=+2640.898919383" watchObservedRunningTime="2025-12-01 09:24:24.997751547 +0000 UTC m=+2640.899860086" Dec 01 09:24:31 crc kubenswrapper[4873]: I1201 09:24:31.061789 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:24:31 crc kubenswrapper[4873]: I1201 09:24:31.077185 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:25:01 crc kubenswrapper[4873]: I1201 09:25:01.059058 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:25:01 crc kubenswrapper[4873]: I1201 09:25:01.059787 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:25:31 crc kubenswrapper[4873]: I1201 09:25:31.060848 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:25:31 crc kubenswrapper[4873]: I1201 09:25:31.061536 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:25:31 crc kubenswrapper[4873]: I1201 09:25:31.061590 4873 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" Dec 01 09:25:31 crc kubenswrapper[4873]: I1201 09:25:31.063421 4873 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5086f7da0c61e3334bc2085e8599b53512aec9e3eb10321166ea8d6e4ebb4c9d"} pod="openshift-machine-config-operator/machine-config-daemon-scwpp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 09:25:31 crc kubenswrapper[4873]: I1201 09:25:31.063690 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" containerID="cri-o://5086f7da0c61e3334bc2085e8599b53512aec9e3eb10321166ea8d6e4ebb4c9d" gracePeriod=600 Dec 01 09:25:31 crc kubenswrapper[4873]: E1201 09:25:31.309288 4873 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfef7b114_0e07_402d_a37b_315c36011f4b.slice/crio-conmon-5086f7da0c61e3334bc2085e8599b53512aec9e3eb10321166ea8d6e4ebb4c9d.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfef7b114_0e07_402d_a37b_315c36011f4b.slice/crio-5086f7da0c61e3334bc2085e8599b53512aec9e3eb10321166ea8d6e4ebb4c9d.scope\": RecentStats: unable to find data in memory cache]" Dec 01 09:25:31 crc kubenswrapper[4873]: I1201 09:25:31.599340 4873 generic.go:334] "Generic (PLEG): container finished" podID="fef7b114-0e07-402d-a37b-315c36011f4b" containerID="5086f7da0c61e3334bc2085e8599b53512aec9e3eb10321166ea8d6e4ebb4c9d" exitCode=0 Dec 01 09:25:31 crc kubenswrapper[4873]: I1201 09:25:31.599399 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" event={"ID":"fef7b114-0e07-402d-a37b-315c36011f4b","Type":"ContainerDied","Data":"5086f7da0c61e3334bc2085e8599b53512aec9e3eb10321166ea8d6e4ebb4c9d"} Dec 01 09:25:31 crc kubenswrapper[4873]: I1201 09:25:31.599889 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" event={"ID":"fef7b114-0e07-402d-a37b-315c36011f4b","Type":"ContainerStarted","Data":"1eb6abed9ec75ad80889e13d075ac04aa13db51de45fa2f980efa98223444319"} Dec 01 09:25:31 crc kubenswrapper[4873]: I1201 09:25:31.599924 4873 scope.go:117] "RemoveContainer" containerID="1a1db8f8327e8cf6dbb94b57ea898b250339bb198ae1ba595cb1d79805fa0db1" Dec 01 09:26:53 crc kubenswrapper[4873]: I1201 09:26:53.805968 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nh4v5"] Dec 01 09:26:53 crc kubenswrapper[4873]: I1201 09:26:53.839752 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nh4v5" Dec 01 09:26:53 crc kubenswrapper[4873]: I1201 09:26:53.844134 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nh4v5"] Dec 01 09:26:53 crc kubenswrapper[4873]: I1201 09:26:53.930655 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d8cd296-b1bd-4923-8906-6b12a81b7967-utilities\") pod \"redhat-marketplace-nh4v5\" (UID: \"4d8cd296-b1bd-4923-8906-6b12a81b7967\") " pod="openshift-marketplace/redhat-marketplace-nh4v5" Dec 01 09:26:53 crc kubenswrapper[4873]: I1201 09:26:53.930728 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxnq6\" (UniqueName: \"kubernetes.io/projected/4d8cd296-b1bd-4923-8906-6b12a81b7967-kube-api-access-lxnq6\") pod \"redhat-marketplace-nh4v5\" (UID: \"4d8cd296-b1bd-4923-8906-6b12a81b7967\") " pod="openshift-marketplace/redhat-marketplace-nh4v5" Dec 01 09:26:53 crc kubenswrapper[4873]: I1201 09:26:53.930861 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d8cd296-b1bd-4923-8906-6b12a81b7967-catalog-content\") pod \"redhat-marketplace-nh4v5\" (UID: \"4d8cd296-b1bd-4923-8906-6b12a81b7967\") " pod="openshift-marketplace/redhat-marketplace-nh4v5" Dec 01 09:26:54 crc kubenswrapper[4873]: I1201 09:26:54.033408 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d8cd296-b1bd-4923-8906-6b12a81b7967-catalog-content\") pod \"redhat-marketplace-nh4v5\" (UID: \"4d8cd296-b1bd-4923-8906-6b12a81b7967\") " pod="openshift-marketplace/redhat-marketplace-nh4v5" Dec 01 09:26:54 crc kubenswrapper[4873]: I1201 09:26:54.033613 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d8cd296-b1bd-4923-8906-6b12a81b7967-utilities\") pod \"redhat-marketplace-nh4v5\" (UID: \"4d8cd296-b1bd-4923-8906-6b12a81b7967\") " pod="openshift-marketplace/redhat-marketplace-nh4v5" Dec 01 09:26:54 crc kubenswrapper[4873]: I1201 09:26:54.033645 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxnq6\" (UniqueName: \"kubernetes.io/projected/4d8cd296-b1bd-4923-8906-6b12a81b7967-kube-api-access-lxnq6\") pod \"redhat-marketplace-nh4v5\" (UID: \"4d8cd296-b1bd-4923-8906-6b12a81b7967\") " pod="openshift-marketplace/redhat-marketplace-nh4v5" Dec 01 09:26:54 crc kubenswrapper[4873]: I1201 09:26:54.034281 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d8cd296-b1bd-4923-8906-6b12a81b7967-catalog-content\") pod \"redhat-marketplace-nh4v5\" (UID: \"4d8cd296-b1bd-4923-8906-6b12a81b7967\") " pod="openshift-marketplace/redhat-marketplace-nh4v5" Dec 01 09:26:54 crc kubenswrapper[4873]: I1201 09:26:54.034435 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d8cd296-b1bd-4923-8906-6b12a81b7967-utilities\") pod \"redhat-marketplace-nh4v5\" (UID: \"4d8cd296-b1bd-4923-8906-6b12a81b7967\") " pod="openshift-marketplace/redhat-marketplace-nh4v5" Dec 01 09:26:54 crc kubenswrapper[4873]: I1201 09:26:54.056235 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxnq6\" (UniqueName: \"kubernetes.io/projected/4d8cd296-b1bd-4923-8906-6b12a81b7967-kube-api-access-lxnq6\") pod \"redhat-marketplace-nh4v5\" (UID: \"4d8cd296-b1bd-4923-8906-6b12a81b7967\") " pod="openshift-marketplace/redhat-marketplace-nh4v5" Dec 01 09:26:54 crc kubenswrapper[4873]: I1201 09:26:54.180888 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nh4v5" Dec 01 09:26:54 crc kubenswrapper[4873]: I1201 09:26:54.698204 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nh4v5"] Dec 01 09:26:55 crc kubenswrapper[4873]: I1201 09:26:55.548430 4873 generic.go:334] "Generic (PLEG): container finished" podID="4d8cd296-b1bd-4923-8906-6b12a81b7967" containerID="5a0350846bdfadb6a18e676ccba9d4e7be3bea3539e28d317fa2307aabac5cc6" exitCode=0 Dec 01 09:26:55 crc kubenswrapper[4873]: I1201 09:26:55.548554 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nh4v5" event={"ID":"4d8cd296-b1bd-4923-8906-6b12a81b7967","Type":"ContainerDied","Data":"5a0350846bdfadb6a18e676ccba9d4e7be3bea3539e28d317fa2307aabac5cc6"} Dec 01 09:26:55 crc kubenswrapper[4873]: I1201 09:26:55.548890 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nh4v5" event={"ID":"4d8cd296-b1bd-4923-8906-6b12a81b7967","Type":"ContainerStarted","Data":"5c14667ae1c0fda84d1f0679beae18d8d5ec4e35a241e3ee5030c2b8eea76e82"} Dec 01 09:26:55 crc kubenswrapper[4873]: I1201 09:26:55.562122 4873 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 09:26:55 crc kubenswrapper[4873]: I1201 09:26:55.611745 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jn8rx"] Dec 01 09:26:55 crc kubenswrapper[4873]: I1201 09:26:55.614145 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jn8rx" Dec 01 09:26:55 crc kubenswrapper[4873]: I1201 09:26:55.624175 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jn8rx"] Dec 01 09:26:55 crc kubenswrapper[4873]: I1201 09:26:55.674413 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zb9rv\" (UniqueName: \"kubernetes.io/projected/9dbdce81-83cc-42f1-811e-a31c02e41f95-kube-api-access-zb9rv\") pod \"redhat-operators-jn8rx\" (UID: \"9dbdce81-83cc-42f1-811e-a31c02e41f95\") " pod="openshift-marketplace/redhat-operators-jn8rx" Dec 01 09:26:55 crc kubenswrapper[4873]: I1201 09:26:55.674533 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9dbdce81-83cc-42f1-811e-a31c02e41f95-catalog-content\") pod \"redhat-operators-jn8rx\" (UID: \"9dbdce81-83cc-42f1-811e-a31c02e41f95\") " pod="openshift-marketplace/redhat-operators-jn8rx" Dec 01 09:26:55 crc kubenswrapper[4873]: I1201 09:26:55.674572 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9dbdce81-83cc-42f1-811e-a31c02e41f95-utilities\") pod \"redhat-operators-jn8rx\" (UID: \"9dbdce81-83cc-42f1-811e-a31c02e41f95\") " pod="openshift-marketplace/redhat-operators-jn8rx" Dec 01 09:26:55 crc kubenswrapper[4873]: I1201 09:26:55.777228 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zb9rv\" (UniqueName: \"kubernetes.io/projected/9dbdce81-83cc-42f1-811e-a31c02e41f95-kube-api-access-zb9rv\") pod \"redhat-operators-jn8rx\" (UID: \"9dbdce81-83cc-42f1-811e-a31c02e41f95\") " pod="openshift-marketplace/redhat-operators-jn8rx" Dec 01 09:26:55 crc kubenswrapper[4873]: I1201 09:26:55.777916 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9dbdce81-83cc-42f1-811e-a31c02e41f95-catalog-content\") pod \"redhat-operators-jn8rx\" (UID: \"9dbdce81-83cc-42f1-811e-a31c02e41f95\") " pod="openshift-marketplace/redhat-operators-jn8rx" Dec 01 09:26:55 crc kubenswrapper[4873]: I1201 09:26:55.777965 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9dbdce81-83cc-42f1-811e-a31c02e41f95-utilities\") pod \"redhat-operators-jn8rx\" (UID: \"9dbdce81-83cc-42f1-811e-a31c02e41f95\") " pod="openshift-marketplace/redhat-operators-jn8rx" Dec 01 09:26:55 crc kubenswrapper[4873]: I1201 09:26:55.778546 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9dbdce81-83cc-42f1-811e-a31c02e41f95-catalog-content\") pod \"redhat-operators-jn8rx\" (UID: \"9dbdce81-83cc-42f1-811e-a31c02e41f95\") " pod="openshift-marketplace/redhat-operators-jn8rx" Dec 01 09:26:55 crc kubenswrapper[4873]: I1201 09:26:55.778669 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9dbdce81-83cc-42f1-811e-a31c02e41f95-utilities\") pod \"redhat-operators-jn8rx\" (UID: \"9dbdce81-83cc-42f1-811e-a31c02e41f95\") " pod="openshift-marketplace/redhat-operators-jn8rx" Dec 01 09:26:55 crc kubenswrapper[4873]: I1201 09:26:55.806912 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zb9rv\" (UniqueName: \"kubernetes.io/projected/9dbdce81-83cc-42f1-811e-a31c02e41f95-kube-api-access-zb9rv\") pod \"redhat-operators-jn8rx\" (UID: \"9dbdce81-83cc-42f1-811e-a31c02e41f95\") " pod="openshift-marketplace/redhat-operators-jn8rx" Dec 01 09:26:55 crc kubenswrapper[4873]: I1201 09:26:55.941607 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jn8rx" Dec 01 09:26:56 crc kubenswrapper[4873]: I1201 09:26:56.445272 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jn8rx"] Dec 01 09:26:56 crc kubenswrapper[4873]: W1201 09:26:56.570677 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9dbdce81_83cc_42f1_811e_a31c02e41f95.slice/crio-d5b3a0f8f72304e51a237e408585e79782a59b4c98bac4ccd2caae679f28e289 WatchSource:0}: Error finding container d5b3a0f8f72304e51a237e408585e79782a59b4c98bac4ccd2caae679f28e289: Status 404 returned error can't find the container with id d5b3a0f8f72304e51a237e408585e79782a59b4c98bac4ccd2caae679f28e289 Dec 01 09:26:57 crc kubenswrapper[4873]: I1201 09:26:57.579937 4873 generic.go:334] "Generic (PLEG): container finished" podID="9dbdce81-83cc-42f1-811e-a31c02e41f95" containerID="548b468f93d2e1d52a61fc3b6515f60c33b5dc9eec140de50660a0e9a4897bec" exitCode=0 Dec 01 09:26:57 crc kubenswrapper[4873]: I1201 09:26:57.580188 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jn8rx" event={"ID":"9dbdce81-83cc-42f1-811e-a31c02e41f95","Type":"ContainerDied","Data":"548b468f93d2e1d52a61fc3b6515f60c33b5dc9eec140de50660a0e9a4897bec"} Dec 01 09:26:57 crc kubenswrapper[4873]: I1201 09:26:57.580545 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jn8rx" event={"ID":"9dbdce81-83cc-42f1-811e-a31c02e41f95","Type":"ContainerStarted","Data":"d5b3a0f8f72304e51a237e408585e79782a59b4c98bac4ccd2caae679f28e289"} Dec 01 09:26:57 crc kubenswrapper[4873]: I1201 09:26:57.587609 4873 generic.go:334] "Generic (PLEG): container finished" podID="4d8cd296-b1bd-4923-8906-6b12a81b7967" containerID="84b79fee0a3afe9a157290d9e11ba39258bdd7bf2b2d14c767f04b9af0917994" exitCode=0 Dec 01 09:26:57 crc kubenswrapper[4873]: I1201 09:26:57.587679 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nh4v5" event={"ID":"4d8cd296-b1bd-4923-8906-6b12a81b7967","Type":"ContainerDied","Data":"84b79fee0a3afe9a157290d9e11ba39258bdd7bf2b2d14c767f04b9af0917994"} Dec 01 09:26:58 crc kubenswrapper[4873]: I1201 09:26:58.602411 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nh4v5" event={"ID":"4d8cd296-b1bd-4923-8906-6b12a81b7967","Type":"ContainerStarted","Data":"4d8a47da9139e3bc8b05985f11a9cc5ef630a57439c26cb07f5d5c7d26a28f82"} Dec 01 09:26:58 crc kubenswrapper[4873]: I1201 09:26:58.627611 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nh4v5" podStartSLOduration=2.994716058 podStartE2EDuration="5.627585393s" podCreationTimestamp="2025-12-01 09:26:53 +0000 UTC" firstStartedPulling="2025-12-01 09:26:55.554285917 +0000 UTC m=+2791.456394456" lastFinishedPulling="2025-12-01 09:26:58.187155252 +0000 UTC m=+2794.089263791" observedRunningTime="2025-12-01 09:26:58.6226028 +0000 UTC m=+2794.524711339" watchObservedRunningTime="2025-12-01 09:26:58.627585393 +0000 UTC m=+2794.529693932" Dec 01 09:26:59 crc kubenswrapper[4873]: I1201 09:26:59.615072 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jn8rx" event={"ID":"9dbdce81-83cc-42f1-811e-a31c02e41f95","Type":"ContainerStarted","Data":"9da4c6bdf522b8e86b7293191771a8162fb8951e899b902975c8bff46852e40f"} Dec 01 09:27:01 crc kubenswrapper[4873]: I1201 09:27:01.636592 4873 generic.go:334] "Generic (PLEG): container finished" podID="9dbdce81-83cc-42f1-811e-a31c02e41f95" containerID="9da4c6bdf522b8e86b7293191771a8162fb8951e899b902975c8bff46852e40f" exitCode=0 Dec 01 09:27:01 crc kubenswrapper[4873]: I1201 09:27:01.636669 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jn8rx" event={"ID":"9dbdce81-83cc-42f1-811e-a31c02e41f95","Type":"ContainerDied","Data":"9da4c6bdf522b8e86b7293191771a8162fb8951e899b902975c8bff46852e40f"} Dec 01 09:27:03 crc kubenswrapper[4873]: I1201 09:27:03.664521 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jn8rx" event={"ID":"9dbdce81-83cc-42f1-811e-a31c02e41f95","Type":"ContainerStarted","Data":"cf51f5c5c9d78aff0ead7ac862493c32c3e99ffe2bcc7f569dcd3d39e2977433"} Dec 01 09:27:03 crc kubenswrapper[4873]: I1201 09:27:03.761174 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jn8rx" podStartSLOduration=3.687735412 podStartE2EDuration="8.761135273s" podCreationTimestamp="2025-12-01 09:26:55 +0000 UTC" firstStartedPulling="2025-12-01 09:26:57.583899361 +0000 UTC m=+2793.486007900" lastFinishedPulling="2025-12-01 09:27:02.657299222 +0000 UTC m=+2798.559407761" observedRunningTime="2025-12-01 09:27:03.749857984 +0000 UTC m=+2799.651966533" watchObservedRunningTime="2025-12-01 09:27:03.761135273 +0000 UTC m=+2799.663243812" Dec 01 09:27:04 crc kubenswrapper[4873]: I1201 09:27:04.181942 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nh4v5" Dec 01 09:27:04 crc kubenswrapper[4873]: I1201 09:27:04.182458 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nh4v5" Dec 01 09:27:04 crc kubenswrapper[4873]: I1201 09:27:04.240710 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nh4v5" Dec 01 09:27:04 crc kubenswrapper[4873]: I1201 09:27:04.728537 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nh4v5" Dec 01 09:27:04 crc kubenswrapper[4873]: I1201 09:27:04.979378 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nh4v5"] Dec 01 09:27:05 crc kubenswrapper[4873]: I1201 09:27:05.941734 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jn8rx" Dec 01 09:27:05 crc kubenswrapper[4873]: I1201 09:27:05.942970 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jn8rx" Dec 01 09:27:06 crc kubenswrapper[4873]: I1201 09:27:06.692884 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nh4v5" podUID="4d8cd296-b1bd-4923-8906-6b12a81b7967" containerName="registry-server" containerID="cri-o://4d8a47da9139e3bc8b05985f11a9cc5ef630a57439c26cb07f5d5c7d26a28f82" gracePeriod=2 Dec 01 09:27:07 crc kubenswrapper[4873]: I1201 09:27:07.004534 4873 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jn8rx" podUID="9dbdce81-83cc-42f1-811e-a31c02e41f95" containerName="registry-server" probeResult="failure" output=< Dec 01 09:27:07 crc kubenswrapper[4873]: timeout: failed to connect service ":50051" within 1s Dec 01 09:27:07 crc kubenswrapper[4873]: > Dec 01 09:27:07 crc kubenswrapper[4873]: I1201 09:27:07.189542 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nh4v5" Dec 01 09:27:07 crc kubenswrapper[4873]: I1201 09:27:07.342517 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d8cd296-b1bd-4923-8906-6b12a81b7967-utilities\") pod \"4d8cd296-b1bd-4923-8906-6b12a81b7967\" (UID: \"4d8cd296-b1bd-4923-8906-6b12a81b7967\") " Dec 01 09:27:07 crc kubenswrapper[4873]: I1201 09:27:07.342616 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d8cd296-b1bd-4923-8906-6b12a81b7967-catalog-content\") pod \"4d8cd296-b1bd-4923-8906-6b12a81b7967\" (UID: \"4d8cd296-b1bd-4923-8906-6b12a81b7967\") " Dec 01 09:27:07 crc kubenswrapper[4873]: I1201 09:27:07.343065 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxnq6\" (UniqueName: \"kubernetes.io/projected/4d8cd296-b1bd-4923-8906-6b12a81b7967-kube-api-access-lxnq6\") pod \"4d8cd296-b1bd-4923-8906-6b12a81b7967\" (UID: \"4d8cd296-b1bd-4923-8906-6b12a81b7967\") " Dec 01 09:27:07 crc kubenswrapper[4873]: I1201 09:27:07.346069 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d8cd296-b1bd-4923-8906-6b12a81b7967-utilities" (OuterVolumeSpecName: "utilities") pod "4d8cd296-b1bd-4923-8906-6b12a81b7967" (UID: "4d8cd296-b1bd-4923-8906-6b12a81b7967"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:27:07 crc kubenswrapper[4873]: I1201 09:27:07.351444 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d8cd296-b1bd-4923-8906-6b12a81b7967-kube-api-access-lxnq6" (OuterVolumeSpecName: "kube-api-access-lxnq6") pod "4d8cd296-b1bd-4923-8906-6b12a81b7967" (UID: "4d8cd296-b1bd-4923-8906-6b12a81b7967"). InnerVolumeSpecName "kube-api-access-lxnq6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:27:07 crc kubenswrapper[4873]: I1201 09:27:07.374053 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d8cd296-b1bd-4923-8906-6b12a81b7967-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4d8cd296-b1bd-4923-8906-6b12a81b7967" (UID: "4d8cd296-b1bd-4923-8906-6b12a81b7967"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:27:07 crc kubenswrapper[4873]: I1201 09:27:07.446636 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxnq6\" (UniqueName: \"kubernetes.io/projected/4d8cd296-b1bd-4923-8906-6b12a81b7967-kube-api-access-lxnq6\") on node \"crc\" DevicePath \"\"" Dec 01 09:27:07 crc kubenswrapper[4873]: I1201 09:27:07.447015 4873 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d8cd296-b1bd-4923-8906-6b12a81b7967-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 09:27:07 crc kubenswrapper[4873]: I1201 09:27:07.447122 4873 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d8cd296-b1bd-4923-8906-6b12a81b7967-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 09:27:07 crc kubenswrapper[4873]: I1201 09:27:07.706199 4873 generic.go:334] "Generic (PLEG): container finished" podID="4d8cd296-b1bd-4923-8906-6b12a81b7967" containerID="4d8a47da9139e3bc8b05985f11a9cc5ef630a57439c26cb07f5d5c7d26a28f82" exitCode=0 Dec 01 09:27:07 crc kubenswrapper[4873]: I1201 09:27:07.706257 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nh4v5" event={"ID":"4d8cd296-b1bd-4923-8906-6b12a81b7967","Type":"ContainerDied","Data":"4d8a47da9139e3bc8b05985f11a9cc5ef630a57439c26cb07f5d5c7d26a28f82"} Dec 01 09:27:07 crc kubenswrapper[4873]: I1201 09:27:07.706294 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nh4v5" event={"ID":"4d8cd296-b1bd-4923-8906-6b12a81b7967","Type":"ContainerDied","Data":"5c14667ae1c0fda84d1f0679beae18d8d5ec4e35a241e3ee5030c2b8eea76e82"} Dec 01 09:27:07 crc kubenswrapper[4873]: I1201 09:27:07.706317 4873 scope.go:117] "RemoveContainer" containerID="4d8a47da9139e3bc8b05985f11a9cc5ef630a57439c26cb07f5d5c7d26a28f82" Dec 01 09:27:07 crc kubenswrapper[4873]: I1201 09:27:07.706422 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nh4v5" Dec 01 09:27:07 crc kubenswrapper[4873]: I1201 09:27:07.750371 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nh4v5"] Dec 01 09:27:07 crc kubenswrapper[4873]: I1201 09:27:07.759311 4873 scope.go:117] "RemoveContainer" containerID="84b79fee0a3afe9a157290d9e11ba39258bdd7bf2b2d14c767f04b9af0917994" Dec 01 09:27:07 crc kubenswrapper[4873]: I1201 09:27:07.765199 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nh4v5"] Dec 01 09:27:07 crc kubenswrapper[4873]: I1201 09:27:07.795905 4873 scope.go:117] "RemoveContainer" containerID="5a0350846bdfadb6a18e676ccba9d4e7be3bea3539e28d317fa2307aabac5cc6" Dec 01 09:27:07 crc kubenswrapper[4873]: I1201 09:27:07.844968 4873 scope.go:117] "RemoveContainer" containerID="4d8a47da9139e3bc8b05985f11a9cc5ef630a57439c26cb07f5d5c7d26a28f82" Dec 01 09:27:07 crc kubenswrapper[4873]: E1201 09:27:07.845477 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d8a47da9139e3bc8b05985f11a9cc5ef630a57439c26cb07f5d5c7d26a28f82\": container with ID starting with 4d8a47da9139e3bc8b05985f11a9cc5ef630a57439c26cb07f5d5c7d26a28f82 not found: ID does not exist" containerID="4d8a47da9139e3bc8b05985f11a9cc5ef630a57439c26cb07f5d5c7d26a28f82" Dec 01 09:27:07 crc kubenswrapper[4873]: I1201 09:27:07.845508 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d8a47da9139e3bc8b05985f11a9cc5ef630a57439c26cb07f5d5c7d26a28f82"} err="failed to get container status \"4d8a47da9139e3bc8b05985f11a9cc5ef630a57439c26cb07f5d5c7d26a28f82\": rpc error: code = NotFound desc = could not find container \"4d8a47da9139e3bc8b05985f11a9cc5ef630a57439c26cb07f5d5c7d26a28f82\": container with ID starting with 4d8a47da9139e3bc8b05985f11a9cc5ef630a57439c26cb07f5d5c7d26a28f82 not found: ID does not exist" Dec 01 09:27:07 crc kubenswrapper[4873]: I1201 09:27:07.845533 4873 scope.go:117] "RemoveContainer" containerID="84b79fee0a3afe9a157290d9e11ba39258bdd7bf2b2d14c767f04b9af0917994" Dec 01 09:27:07 crc kubenswrapper[4873]: E1201 09:27:07.845771 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84b79fee0a3afe9a157290d9e11ba39258bdd7bf2b2d14c767f04b9af0917994\": container with ID starting with 84b79fee0a3afe9a157290d9e11ba39258bdd7bf2b2d14c767f04b9af0917994 not found: ID does not exist" containerID="84b79fee0a3afe9a157290d9e11ba39258bdd7bf2b2d14c767f04b9af0917994" Dec 01 09:27:07 crc kubenswrapper[4873]: I1201 09:27:07.845802 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84b79fee0a3afe9a157290d9e11ba39258bdd7bf2b2d14c767f04b9af0917994"} err="failed to get container status \"84b79fee0a3afe9a157290d9e11ba39258bdd7bf2b2d14c767f04b9af0917994\": rpc error: code = NotFound desc = could not find container \"84b79fee0a3afe9a157290d9e11ba39258bdd7bf2b2d14c767f04b9af0917994\": container with ID starting with 84b79fee0a3afe9a157290d9e11ba39258bdd7bf2b2d14c767f04b9af0917994 not found: ID does not exist" Dec 01 09:27:07 crc kubenswrapper[4873]: I1201 09:27:07.845836 4873 scope.go:117] "RemoveContainer" containerID="5a0350846bdfadb6a18e676ccba9d4e7be3bea3539e28d317fa2307aabac5cc6" Dec 01 09:27:07 crc kubenswrapper[4873]: E1201 09:27:07.846267 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a0350846bdfadb6a18e676ccba9d4e7be3bea3539e28d317fa2307aabac5cc6\": container with ID starting with 5a0350846bdfadb6a18e676ccba9d4e7be3bea3539e28d317fa2307aabac5cc6 not found: ID does not exist" containerID="5a0350846bdfadb6a18e676ccba9d4e7be3bea3539e28d317fa2307aabac5cc6" Dec 01 09:27:07 crc kubenswrapper[4873]: I1201 09:27:07.846300 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a0350846bdfadb6a18e676ccba9d4e7be3bea3539e28d317fa2307aabac5cc6"} err="failed to get container status \"5a0350846bdfadb6a18e676ccba9d4e7be3bea3539e28d317fa2307aabac5cc6\": rpc error: code = NotFound desc = could not find container \"5a0350846bdfadb6a18e676ccba9d4e7be3bea3539e28d317fa2307aabac5cc6\": container with ID starting with 5a0350846bdfadb6a18e676ccba9d4e7be3bea3539e28d317fa2307aabac5cc6 not found: ID does not exist" Dec 01 09:27:08 crc kubenswrapper[4873]: I1201 09:27:08.442743 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d8cd296-b1bd-4923-8906-6b12a81b7967" path="/var/lib/kubelet/pods/4d8cd296-b1bd-4923-8906-6b12a81b7967/volumes" Dec 01 09:27:16 crc kubenswrapper[4873]: I1201 09:27:16.000267 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jn8rx" Dec 01 09:27:16 crc kubenswrapper[4873]: I1201 09:27:16.062144 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jn8rx" Dec 01 09:27:16 crc kubenswrapper[4873]: I1201 09:27:16.251015 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jn8rx"] Dec 01 09:27:17 crc kubenswrapper[4873]: I1201 09:27:17.814018 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jn8rx" podUID="9dbdce81-83cc-42f1-811e-a31c02e41f95" containerName="registry-server" containerID="cri-o://cf51f5c5c9d78aff0ead7ac862493c32c3e99ffe2bcc7f569dcd3d39e2977433" gracePeriod=2 Dec 01 09:27:18 crc kubenswrapper[4873]: I1201 09:27:18.304084 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jn8rx" Dec 01 09:27:18 crc kubenswrapper[4873]: I1201 09:27:18.323224 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9dbdce81-83cc-42f1-811e-a31c02e41f95-utilities\") pod \"9dbdce81-83cc-42f1-811e-a31c02e41f95\" (UID: \"9dbdce81-83cc-42f1-811e-a31c02e41f95\") " Dec 01 09:27:18 crc kubenswrapper[4873]: I1201 09:27:18.323360 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zb9rv\" (UniqueName: \"kubernetes.io/projected/9dbdce81-83cc-42f1-811e-a31c02e41f95-kube-api-access-zb9rv\") pod \"9dbdce81-83cc-42f1-811e-a31c02e41f95\" (UID: \"9dbdce81-83cc-42f1-811e-a31c02e41f95\") " Dec 01 09:27:18 crc kubenswrapper[4873]: I1201 09:27:18.323719 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9dbdce81-83cc-42f1-811e-a31c02e41f95-catalog-content\") pod \"9dbdce81-83cc-42f1-811e-a31c02e41f95\" (UID: \"9dbdce81-83cc-42f1-811e-a31c02e41f95\") " Dec 01 09:27:18 crc kubenswrapper[4873]: I1201 09:27:18.332007 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9dbdce81-83cc-42f1-811e-a31c02e41f95-kube-api-access-zb9rv" (OuterVolumeSpecName: "kube-api-access-zb9rv") pod "9dbdce81-83cc-42f1-811e-a31c02e41f95" (UID: "9dbdce81-83cc-42f1-811e-a31c02e41f95"). InnerVolumeSpecName "kube-api-access-zb9rv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:27:18 crc kubenswrapper[4873]: I1201 09:27:18.338556 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9dbdce81-83cc-42f1-811e-a31c02e41f95-utilities" (OuterVolumeSpecName: "utilities") pod "9dbdce81-83cc-42f1-811e-a31c02e41f95" (UID: "9dbdce81-83cc-42f1-811e-a31c02e41f95"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:27:18 crc kubenswrapper[4873]: I1201 09:27:18.444387 4873 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9dbdce81-83cc-42f1-811e-a31c02e41f95-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 09:27:18 crc kubenswrapper[4873]: I1201 09:27:18.444441 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zb9rv\" (UniqueName: \"kubernetes.io/projected/9dbdce81-83cc-42f1-811e-a31c02e41f95-kube-api-access-zb9rv\") on node \"crc\" DevicePath \"\"" Dec 01 09:27:18 crc kubenswrapper[4873]: I1201 09:27:18.451592 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9dbdce81-83cc-42f1-811e-a31c02e41f95-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9dbdce81-83cc-42f1-811e-a31c02e41f95" (UID: "9dbdce81-83cc-42f1-811e-a31c02e41f95"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:27:18 crc kubenswrapper[4873]: I1201 09:27:18.546802 4873 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9dbdce81-83cc-42f1-811e-a31c02e41f95-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 09:27:18 crc kubenswrapper[4873]: I1201 09:27:18.826840 4873 generic.go:334] "Generic (PLEG): container finished" podID="9dbdce81-83cc-42f1-811e-a31c02e41f95" containerID="cf51f5c5c9d78aff0ead7ac862493c32c3e99ffe2bcc7f569dcd3d39e2977433" exitCode=0 Dec 01 09:27:18 crc kubenswrapper[4873]: I1201 09:27:18.826948 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jn8rx" Dec 01 09:27:18 crc kubenswrapper[4873]: I1201 09:27:18.826927 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jn8rx" event={"ID":"9dbdce81-83cc-42f1-811e-a31c02e41f95","Type":"ContainerDied","Data":"cf51f5c5c9d78aff0ead7ac862493c32c3e99ffe2bcc7f569dcd3d39e2977433"} Dec 01 09:27:18 crc kubenswrapper[4873]: I1201 09:27:18.827092 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jn8rx" event={"ID":"9dbdce81-83cc-42f1-811e-a31c02e41f95","Type":"ContainerDied","Data":"d5b3a0f8f72304e51a237e408585e79782a59b4c98bac4ccd2caae679f28e289"} Dec 01 09:27:18 crc kubenswrapper[4873]: I1201 09:27:18.827126 4873 scope.go:117] "RemoveContainer" containerID="cf51f5c5c9d78aff0ead7ac862493c32c3e99ffe2bcc7f569dcd3d39e2977433" Dec 01 09:27:18 crc kubenswrapper[4873]: I1201 09:27:18.859726 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jn8rx"] Dec 01 09:27:18 crc kubenswrapper[4873]: I1201 09:27:18.860057 4873 scope.go:117] "RemoveContainer" containerID="9da4c6bdf522b8e86b7293191771a8162fb8951e899b902975c8bff46852e40f" Dec 01 09:27:18 crc kubenswrapper[4873]: I1201 09:27:18.869243 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jn8rx"] Dec 01 09:27:18 crc kubenswrapper[4873]: I1201 09:27:18.893071 4873 scope.go:117] "RemoveContainer" containerID="548b468f93d2e1d52a61fc3b6515f60c33b5dc9eec140de50660a0e9a4897bec" Dec 01 09:27:18 crc kubenswrapper[4873]: I1201 09:27:18.936945 4873 scope.go:117] "RemoveContainer" containerID="cf51f5c5c9d78aff0ead7ac862493c32c3e99ffe2bcc7f569dcd3d39e2977433" Dec 01 09:27:18 crc kubenswrapper[4873]: E1201 09:27:18.937726 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf51f5c5c9d78aff0ead7ac862493c32c3e99ffe2bcc7f569dcd3d39e2977433\": container with ID starting with cf51f5c5c9d78aff0ead7ac862493c32c3e99ffe2bcc7f569dcd3d39e2977433 not found: ID does not exist" containerID="cf51f5c5c9d78aff0ead7ac862493c32c3e99ffe2bcc7f569dcd3d39e2977433" Dec 01 09:27:18 crc kubenswrapper[4873]: I1201 09:27:18.937796 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf51f5c5c9d78aff0ead7ac862493c32c3e99ffe2bcc7f569dcd3d39e2977433"} err="failed to get container status \"cf51f5c5c9d78aff0ead7ac862493c32c3e99ffe2bcc7f569dcd3d39e2977433\": rpc error: code = NotFound desc = could not find container \"cf51f5c5c9d78aff0ead7ac862493c32c3e99ffe2bcc7f569dcd3d39e2977433\": container with ID starting with cf51f5c5c9d78aff0ead7ac862493c32c3e99ffe2bcc7f569dcd3d39e2977433 not found: ID does not exist" Dec 01 09:27:18 crc kubenswrapper[4873]: I1201 09:27:18.937833 4873 scope.go:117] "RemoveContainer" containerID="9da4c6bdf522b8e86b7293191771a8162fb8951e899b902975c8bff46852e40f" Dec 01 09:27:18 crc kubenswrapper[4873]: E1201 09:27:18.938524 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9da4c6bdf522b8e86b7293191771a8162fb8951e899b902975c8bff46852e40f\": container with ID starting with 9da4c6bdf522b8e86b7293191771a8162fb8951e899b902975c8bff46852e40f not found: ID does not exist" containerID="9da4c6bdf522b8e86b7293191771a8162fb8951e899b902975c8bff46852e40f" Dec 01 09:27:18 crc kubenswrapper[4873]: I1201 09:27:18.938567 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9da4c6bdf522b8e86b7293191771a8162fb8951e899b902975c8bff46852e40f"} err="failed to get container status \"9da4c6bdf522b8e86b7293191771a8162fb8951e899b902975c8bff46852e40f\": rpc error: code = NotFound desc = could not find container \"9da4c6bdf522b8e86b7293191771a8162fb8951e899b902975c8bff46852e40f\": container with ID starting with 9da4c6bdf522b8e86b7293191771a8162fb8951e899b902975c8bff46852e40f not found: ID does not exist" Dec 01 09:27:18 crc kubenswrapper[4873]: I1201 09:27:18.938587 4873 scope.go:117] "RemoveContainer" containerID="548b468f93d2e1d52a61fc3b6515f60c33b5dc9eec140de50660a0e9a4897bec" Dec 01 09:27:18 crc kubenswrapper[4873]: E1201 09:27:18.938856 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"548b468f93d2e1d52a61fc3b6515f60c33b5dc9eec140de50660a0e9a4897bec\": container with ID starting with 548b468f93d2e1d52a61fc3b6515f60c33b5dc9eec140de50660a0e9a4897bec not found: ID does not exist" containerID="548b468f93d2e1d52a61fc3b6515f60c33b5dc9eec140de50660a0e9a4897bec" Dec 01 09:27:18 crc kubenswrapper[4873]: I1201 09:27:18.938889 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"548b468f93d2e1d52a61fc3b6515f60c33b5dc9eec140de50660a0e9a4897bec"} err="failed to get container status \"548b468f93d2e1d52a61fc3b6515f60c33b5dc9eec140de50660a0e9a4897bec\": rpc error: code = NotFound desc = could not find container \"548b468f93d2e1d52a61fc3b6515f60c33b5dc9eec140de50660a0e9a4897bec\": container with ID starting with 548b468f93d2e1d52a61fc3b6515f60c33b5dc9eec140de50660a0e9a4897bec not found: ID does not exist" Dec 01 09:27:20 crc kubenswrapper[4873]: I1201 09:27:20.444305 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9dbdce81-83cc-42f1-811e-a31c02e41f95" path="/var/lib/kubelet/pods/9dbdce81-83cc-42f1-811e-a31c02e41f95/volumes" Dec 01 09:27:31 crc kubenswrapper[4873]: I1201 09:27:31.059441 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:27:31 crc kubenswrapper[4873]: I1201 09:27:31.060086 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:27:37 crc kubenswrapper[4873]: I1201 09:27:37.559259 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7zq5j"] Dec 01 09:27:37 crc kubenswrapper[4873]: E1201 09:27:37.560281 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dbdce81-83cc-42f1-811e-a31c02e41f95" containerName="extract-utilities" Dec 01 09:27:37 crc kubenswrapper[4873]: I1201 09:27:37.560307 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dbdce81-83cc-42f1-811e-a31c02e41f95" containerName="extract-utilities" Dec 01 09:27:37 crc kubenswrapper[4873]: E1201 09:27:37.560363 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dbdce81-83cc-42f1-811e-a31c02e41f95" containerName="registry-server" Dec 01 09:27:37 crc kubenswrapper[4873]: I1201 09:27:37.560372 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dbdce81-83cc-42f1-811e-a31c02e41f95" containerName="registry-server" Dec 01 09:27:37 crc kubenswrapper[4873]: E1201 09:27:37.560383 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d8cd296-b1bd-4923-8906-6b12a81b7967" containerName="extract-content" Dec 01 09:27:37 crc kubenswrapper[4873]: I1201 09:27:37.560390 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d8cd296-b1bd-4923-8906-6b12a81b7967" containerName="extract-content" Dec 01 09:27:37 crc kubenswrapper[4873]: E1201 09:27:37.560400 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d8cd296-b1bd-4923-8906-6b12a81b7967" containerName="extract-utilities" Dec 01 09:27:37 crc kubenswrapper[4873]: I1201 09:27:37.560406 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d8cd296-b1bd-4923-8906-6b12a81b7967" containerName="extract-utilities" Dec 01 09:27:37 crc kubenswrapper[4873]: E1201 09:27:37.560432 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dbdce81-83cc-42f1-811e-a31c02e41f95" containerName="extract-content" Dec 01 09:27:37 crc kubenswrapper[4873]: I1201 09:27:37.560438 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dbdce81-83cc-42f1-811e-a31c02e41f95" containerName="extract-content" Dec 01 09:27:37 crc kubenswrapper[4873]: E1201 09:27:37.560451 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d8cd296-b1bd-4923-8906-6b12a81b7967" containerName="registry-server" Dec 01 09:27:37 crc kubenswrapper[4873]: I1201 09:27:37.560458 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d8cd296-b1bd-4923-8906-6b12a81b7967" containerName="registry-server" Dec 01 09:27:37 crc kubenswrapper[4873]: I1201 09:27:37.560726 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="9dbdce81-83cc-42f1-811e-a31c02e41f95" containerName="registry-server" Dec 01 09:27:37 crc kubenswrapper[4873]: I1201 09:27:37.560753 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d8cd296-b1bd-4923-8906-6b12a81b7967" containerName="registry-server" Dec 01 09:27:37 crc kubenswrapper[4873]: I1201 09:27:37.562507 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7zq5j" Dec 01 09:27:37 crc kubenswrapper[4873]: I1201 09:27:37.575713 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7zq5j"] Dec 01 09:27:37 crc kubenswrapper[4873]: I1201 09:27:37.702240 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3adbfc1d-7e40-4bfc-ab02-bcec5b0a5c38-catalog-content\") pod \"community-operators-7zq5j\" (UID: \"3adbfc1d-7e40-4bfc-ab02-bcec5b0a5c38\") " pod="openshift-marketplace/community-operators-7zq5j" Dec 01 09:27:37 crc kubenswrapper[4873]: I1201 09:27:37.702308 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3adbfc1d-7e40-4bfc-ab02-bcec5b0a5c38-utilities\") pod \"community-operators-7zq5j\" (UID: \"3adbfc1d-7e40-4bfc-ab02-bcec5b0a5c38\") " pod="openshift-marketplace/community-operators-7zq5j" Dec 01 09:27:37 crc kubenswrapper[4873]: I1201 09:27:37.702342 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxcmb\" (UniqueName: \"kubernetes.io/projected/3adbfc1d-7e40-4bfc-ab02-bcec5b0a5c38-kube-api-access-kxcmb\") pod \"community-operators-7zq5j\" (UID: \"3adbfc1d-7e40-4bfc-ab02-bcec5b0a5c38\") " pod="openshift-marketplace/community-operators-7zq5j" Dec 01 09:27:37 crc kubenswrapper[4873]: I1201 09:27:37.804235 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3adbfc1d-7e40-4bfc-ab02-bcec5b0a5c38-catalog-content\") pod \"community-operators-7zq5j\" (UID: \"3adbfc1d-7e40-4bfc-ab02-bcec5b0a5c38\") " pod="openshift-marketplace/community-operators-7zq5j" Dec 01 09:27:37 crc kubenswrapper[4873]: I1201 09:27:37.804321 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3adbfc1d-7e40-4bfc-ab02-bcec5b0a5c38-utilities\") pod \"community-operators-7zq5j\" (UID: \"3adbfc1d-7e40-4bfc-ab02-bcec5b0a5c38\") " pod="openshift-marketplace/community-operators-7zq5j" Dec 01 09:27:37 crc kubenswrapper[4873]: I1201 09:27:37.804348 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxcmb\" (UniqueName: \"kubernetes.io/projected/3adbfc1d-7e40-4bfc-ab02-bcec5b0a5c38-kube-api-access-kxcmb\") pod \"community-operators-7zq5j\" (UID: \"3adbfc1d-7e40-4bfc-ab02-bcec5b0a5c38\") " pod="openshift-marketplace/community-operators-7zq5j" Dec 01 09:27:37 crc kubenswrapper[4873]: I1201 09:27:37.805689 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3adbfc1d-7e40-4bfc-ab02-bcec5b0a5c38-catalog-content\") pod \"community-operators-7zq5j\" (UID: \"3adbfc1d-7e40-4bfc-ab02-bcec5b0a5c38\") " pod="openshift-marketplace/community-operators-7zq5j" Dec 01 09:27:37 crc kubenswrapper[4873]: I1201 09:27:37.805874 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3adbfc1d-7e40-4bfc-ab02-bcec5b0a5c38-utilities\") pod \"community-operators-7zq5j\" (UID: \"3adbfc1d-7e40-4bfc-ab02-bcec5b0a5c38\") " pod="openshift-marketplace/community-operators-7zq5j" Dec 01 09:27:37 crc kubenswrapper[4873]: I1201 09:27:37.829822 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxcmb\" (UniqueName: \"kubernetes.io/projected/3adbfc1d-7e40-4bfc-ab02-bcec5b0a5c38-kube-api-access-kxcmb\") pod \"community-operators-7zq5j\" (UID: \"3adbfc1d-7e40-4bfc-ab02-bcec5b0a5c38\") " pod="openshift-marketplace/community-operators-7zq5j" Dec 01 09:27:37 crc kubenswrapper[4873]: I1201 09:27:37.885551 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7zq5j" Dec 01 09:27:38 crc kubenswrapper[4873]: I1201 09:27:38.578474 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7zq5j"] Dec 01 09:27:39 crc kubenswrapper[4873]: I1201 09:27:39.026813 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7zq5j" event={"ID":"3adbfc1d-7e40-4bfc-ab02-bcec5b0a5c38","Type":"ContainerStarted","Data":"be8c4e6d0d8873c40f784f5fe4a1296aaf95cdf129333ba2276b6b6e83c6c4df"} Dec 01 09:27:40 crc kubenswrapper[4873]: I1201 09:27:40.040310 4873 generic.go:334] "Generic (PLEG): container finished" podID="3adbfc1d-7e40-4bfc-ab02-bcec5b0a5c38" containerID="2d4679997d4f626009ea1e27b46e57c7e6ae0454b66b1c57574b7486c0678750" exitCode=0 Dec 01 09:27:40 crc kubenswrapper[4873]: I1201 09:27:40.040437 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7zq5j" event={"ID":"3adbfc1d-7e40-4bfc-ab02-bcec5b0a5c38","Type":"ContainerDied","Data":"2d4679997d4f626009ea1e27b46e57c7e6ae0454b66b1c57574b7486c0678750"} Dec 01 09:27:46 crc kubenswrapper[4873]: I1201 09:27:46.115408 4873 generic.go:334] "Generic (PLEG): container finished" podID="3adbfc1d-7e40-4bfc-ab02-bcec5b0a5c38" containerID="587d3efe47376c92fd8f1c5670b5f1e7c50a2669d50292038cbbc6283a7ac6f2" exitCode=0 Dec 01 09:27:46 crc kubenswrapper[4873]: I1201 09:27:46.115520 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7zq5j" event={"ID":"3adbfc1d-7e40-4bfc-ab02-bcec5b0a5c38","Type":"ContainerDied","Data":"587d3efe47376c92fd8f1c5670b5f1e7c50a2669d50292038cbbc6283a7ac6f2"} Dec 01 09:27:48 crc kubenswrapper[4873]: I1201 09:27:48.139913 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7zq5j" event={"ID":"3adbfc1d-7e40-4bfc-ab02-bcec5b0a5c38","Type":"ContainerStarted","Data":"f40ba6d6a91209dcd51e773bae913f55abd4218c9d5c0996459b02d082680b82"} Dec 01 09:27:48 crc kubenswrapper[4873]: I1201 09:27:48.164965 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7zq5j" podStartSLOduration=4.012459845 podStartE2EDuration="11.164936693s" podCreationTimestamp="2025-12-01 09:27:37 +0000 UTC" firstStartedPulling="2025-12-01 09:27:40.043313959 +0000 UTC m=+2835.945422498" lastFinishedPulling="2025-12-01 09:27:47.195790807 +0000 UTC m=+2843.097899346" observedRunningTime="2025-12-01 09:27:48.161137379 +0000 UTC m=+2844.063245938" watchObservedRunningTime="2025-12-01 09:27:48.164936693 +0000 UTC m=+2844.067045232" Dec 01 09:27:57 crc kubenswrapper[4873]: I1201 09:27:57.885946 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7zq5j" Dec 01 09:27:57 crc kubenswrapper[4873]: I1201 09:27:57.886791 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7zq5j" Dec 01 09:27:57 crc kubenswrapper[4873]: I1201 09:27:57.972917 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7zq5j" Dec 01 09:27:58 crc kubenswrapper[4873]: I1201 09:27:58.391789 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7zq5j" Dec 01 09:27:58 crc kubenswrapper[4873]: I1201 09:27:58.489701 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7zq5j"] Dec 01 09:27:58 crc kubenswrapper[4873]: I1201 09:27:58.568654 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6rsqz"] Dec 01 09:27:58 crc kubenswrapper[4873]: I1201 09:27:58.569357 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6rsqz" podUID="1c2e9631-829e-4ee3-af26-1f8bffc37029" containerName="registry-server" containerID="cri-o://f165854cd4bc8b7a60aa520b4bdf503536cc30f8ac115ce50673b958ddd9c771" gracePeriod=2 Dec 01 09:28:00 crc kubenswrapper[4873]: I1201 09:28:00.359700 4873 generic.go:334] "Generic (PLEG): container finished" podID="1c2e9631-829e-4ee3-af26-1f8bffc37029" containerID="f165854cd4bc8b7a60aa520b4bdf503536cc30f8ac115ce50673b958ddd9c771" exitCode=0 Dec 01 09:28:00 crc kubenswrapper[4873]: I1201 09:28:00.359780 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6rsqz" event={"ID":"1c2e9631-829e-4ee3-af26-1f8bffc37029","Type":"ContainerDied","Data":"f165854cd4bc8b7a60aa520b4bdf503536cc30f8ac115ce50673b958ddd9c771"} Dec 01 09:28:00 crc kubenswrapper[4873]: I1201 09:28:00.360663 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6rsqz" event={"ID":"1c2e9631-829e-4ee3-af26-1f8bffc37029","Type":"ContainerDied","Data":"fba2a7d144f7bb0c78e1ead6ce3bcc8858cad994e8aa7736d4b39e87fc5f6eda"} Dec 01 09:28:00 crc kubenswrapper[4873]: I1201 09:28:00.360682 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fba2a7d144f7bb0c78e1ead6ce3bcc8858cad994e8aa7736d4b39e87fc5f6eda" Dec 01 09:28:00 crc kubenswrapper[4873]: I1201 09:28:00.409505 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6rsqz" Dec 01 09:28:00 crc kubenswrapper[4873]: I1201 09:28:00.608122 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c2e9631-829e-4ee3-af26-1f8bffc37029-utilities\") pod \"1c2e9631-829e-4ee3-af26-1f8bffc37029\" (UID: \"1c2e9631-829e-4ee3-af26-1f8bffc37029\") " Dec 01 09:28:00 crc kubenswrapper[4873]: I1201 09:28:00.608436 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8s284\" (UniqueName: \"kubernetes.io/projected/1c2e9631-829e-4ee3-af26-1f8bffc37029-kube-api-access-8s284\") pod \"1c2e9631-829e-4ee3-af26-1f8bffc37029\" (UID: \"1c2e9631-829e-4ee3-af26-1f8bffc37029\") " Dec 01 09:28:00 crc kubenswrapper[4873]: I1201 09:28:00.608556 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c2e9631-829e-4ee3-af26-1f8bffc37029-catalog-content\") pod \"1c2e9631-829e-4ee3-af26-1f8bffc37029\" (UID: \"1c2e9631-829e-4ee3-af26-1f8bffc37029\") " Dec 01 09:28:00 crc kubenswrapper[4873]: I1201 09:28:00.610667 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c2e9631-829e-4ee3-af26-1f8bffc37029-utilities" (OuterVolumeSpecName: "utilities") pod "1c2e9631-829e-4ee3-af26-1f8bffc37029" (UID: "1c2e9631-829e-4ee3-af26-1f8bffc37029"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:28:00 crc kubenswrapper[4873]: I1201 09:28:00.631162 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c2e9631-829e-4ee3-af26-1f8bffc37029-kube-api-access-8s284" (OuterVolumeSpecName: "kube-api-access-8s284") pod "1c2e9631-829e-4ee3-af26-1f8bffc37029" (UID: "1c2e9631-829e-4ee3-af26-1f8bffc37029"). InnerVolumeSpecName "kube-api-access-8s284". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:28:00 crc kubenswrapper[4873]: I1201 09:28:00.675407 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c2e9631-829e-4ee3-af26-1f8bffc37029-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1c2e9631-829e-4ee3-af26-1f8bffc37029" (UID: "1c2e9631-829e-4ee3-af26-1f8bffc37029"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:28:00 crc kubenswrapper[4873]: I1201 09:28:00.710871 4873 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c2e9631-829e-4ee3-af26-1f8bffc37029-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 09:28:00 crc kubenswrapper[4873]: I1201 09:28:00.710914 4873 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c2e9631-829e-4ee3-af26-1f8bffc37029-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 09:28:00 crc kubenswrapper[4873]: I1201 09:28:00.710924 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8s284\" (UniqueName: \"kubernetes.io/projected/1c2e9631-829e-4ee3-af26-1f8bffc37029-kube-api-access-8s284\") on node \"crc\" DevicePath \"\"" Dec 01 09:28:01 crc kubenswrapper[4873]: I1201 09:28:01.059679 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:28:01 crc kubenswrapper[4873]: I1201 09:28:01.060320 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:28:01 crc kubenswrapper[4873]: I1201 09:28:01.371232 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6rsqz" Dec 01 09:28:01 crc kubenswrapper[4873]: I1201 09:28:01.414968 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6rsqz"] Dec 01 09:28:01 crc kubenswrapper[4873]: I1201 09:28:01.427250 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6rsqz"] Dec 01 09:28:02 crc kubenswrapper[4873]: I1201 09:28:02.442419 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c2e9631-829e-4ee3-af26-1f8bffc37029" path="/var/lib/kubelet/pods/1c2e9631-829e-4ee3-af26-1f8bffc37029/volumes" Dec 01 09:28:15 crc kubenswrapper[4873]: I1201 09:28:15.430080 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-z85vq"] Dec 01 09:28:15 crc kubenswrapper[4873]: E1201 09:28:15.431423 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c2e9631-829e-4ee3-af26-1f8bffc37029" containerName="registry-server" Dec 01 09:28:15 crc kubenswrapper[4873]: I1201 09:28:15.431442 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c2e9631-829e-4ee3-af26-1f8bffc37029" containerName="registry-server" Dec 01 09:28:15 crc kubenswrapper[4873]: E1201 09:28:15.431482 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c2e9631-829e-4ee3-af26-1f8bffc37029" containerName="extract-content" Dec 01 09:28:15 crc kubenswrapper[4873]: I1201 09:28:15.431488 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c2e9631-829e-4ee3-af26-1f8bffc37029" containerName="extract-content" Dec 01 09:28:15 crc kubenswrapper[4873]: E1201 09:28:15.431510 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c2e9631-829e-4ee3-af26-1f8bffc37029" containerName="extract-utilities" Dec 01 09:28:15 crc kubenswrapper[4873]: I1201 09:28:15.431518 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c2e9631-829e-4ee3-af26-1f8bffc37029" containerName="extract-utilities" Dec 01 09:28:15 crc kubenswrapper[4873]: I1201 09:28:15.431752 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c2e9631-829e-4ee3-af26-1f8bffc37029" containerName="registry-server" Dec 01 09:28:15 crc kubenswrapper[4873]: I1201 09:28:15.433590 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z85vq" Dec 01 09:28:15 crc kubenswrapper[4873]: I1201 09:28:15.441785 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-z85vq"] Dec 01 09:28:15 crc kubenswrapper[4873]: I1201 09:28:15.552801 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7kw9\" (UniqueName: \"kubernetes.io/projected/36b2e93f-f26a-4b05-83f9-3479f92ce038-kube-api-access-r7kw9\") pod \"certified-operators-z85vq\" (UID: \"36b2e93f-f26a-4b05-83f9-3479f92ce038\") " pod="openshift-marketplace/certified-operators-z85vq" Dec 01 09:28:15 crc kubenswrapper[4873]: I1201 09:28:15.552897 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36b2e93f-f26a-4b05-83f9-3479f92ce038-utilities\") pod \"certified-operators-z85vq\" (UID: \"36b2e93f-f26a-4b05-83f9-3479f92ce038\") " pod="openshift-marketplace/certified-operators-z85vq" Dec 01 09:28:15 crc kubenswrapper[4873]: I1201 09:28:15.553241 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36b2e93f-f26a-4b05-83f9-3479f92ce038-catalog-content\") pod \"certified-operators-z85vq\" (UID: \"36b2e93f-f26a-4b05-83f9-3479f92ce038\") " pod="openshift-marketplace/certified-operators-z85vq" Dec 01 09:28:15 crc kubenswrapper[4873]: I1201 09:28:15.654901 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36b2e93f-f26a-4b05-83f9-3479f92ce038-catalog-content\") pod \"certified-operators-z85vq\" (UID: \"36b2e93f-f26a-4b05-83f9-3479f92ce038\") " pod="openshift-marketplace/certified-operators-z85vq" Dec 01 09:28:15 crc kubenswrapper[4873]: I1201 09:28:15.655129 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7kw9\" (UniqueName: \"kubernetes.io/projected/36b2e93f-f26a-4b05-83f9-3479f92ce038-kube-api-access-r7kw9\") pod \"certified-operators-z85vq\" (UID: \"36b2e93f-f26a-4b05-83f9-3479f92ce038\") " pod="openshift-marketplace/certified-operators-z85vq" Dec 01 09:28:15 crc kubenswrapper[4873]: I1201 09:28:15.655192 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36b2e93f-f26a-4b05-83f9-3479f92ce038-utilities\") pod \"certified-operators-z85vq\" (UID: \"36b2e93f-f26a-4b05-83f9-3479f92ce038\") " pod="openshift-marketplace/certified-operators-z85vq" Dec 01 09:28:15 crc kubenswrapper[4873]: I1201 09:28:15.655609 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36b2e93f-f26a-4b05-83f9-3479f92ce038-catalog-content\") pod \"certified-operators-z85vq\" (UID: \"36b2e93f-f26a-4b05-83f9-3479f92ce038\") " pod="openshift-marketplace/certified-operators-z85vq" Dec 01 09:28:15 crc kubenswrapper[4873]: I1201 09:28:15.655744 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36b2e93f-f26a-4b05-83f9-3479f92ce038-utilities\") pod \"certified-operators-z85vq\" (UID: \"36b2e93f-f26a-4b05-83f9-3479f92ce038\") " pod="openshift-marketplace/certified-operators-z85vq" Dec 01 09:28:15 crc kubenswrapper[4873]: I1201 09:28:15.680217 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7kw9\" (UniqueName: \"kubernetes.io/projected/36b2e93f-f26a-4b05-83f9-3479f92ce038-kube-api-access-r7kw9\") pod \"certified-operators-z85vq\" (UID: \"36b2e93f-f26a-4b05-83f9-3479f92ce038\") " pod="openshift-marketplace/certified-operators-z85vq" Dec 01 09:28:15 crc kubenswrapper[4873]: I1201 09:28:15.758816 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z85vq" Dec 01 09:28:16 crc kubenswrapper[4873]: I1201 09:28:16.346580 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-z85vq"] Dec 01 09:28:16 crc kubenswrapper[4873]: I1201 09:28:16.539929 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z85vq" event={"ID":"36b2e93f-f26a-4b05-83f9-3479f92ce038","Type":"ContainerStarted","Data":"0f503c68fd10a4ae393772931169bee60d6194931eaad22416870f38c0070b0d"} Dec 01 09:28:17 crc kubenswrapper[4873]: I1201 09:28:17.553495 4873 generic.go:334] "Generic (PLEG): container finished" podID="36b2e93f-f26a-4b05-83f9-3479f92ce038" containerID="a3aae70def07ef8ca54a1bcbbbee1691d453bd0e07d9bbc469320b4aa49f127f" exitCode=0 Dec 01 09:28:17 crc kubenswrapper[4873]: I1201 09:28:17.553556 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z85vq" event={"ID":"36b2e93f-f26a-4b05-83f9-3479f92ce038","Type":"ContainerDied","Data":"a3aae70def07ef8ca54a1bcbbbee1691d453bd0e07d9bbc469320b4aa49f127f"} Dec 01 09:28:19 crc kubenswrapper[4873]: I1201 09:28:19.578903 4873 generic.go:334] "Generic (PLEG): container finished" podID="36b2e93f-f26a-4b05-83f9-3479f92ce038" containerID="14fb93271d6a7e20e0456a629c4a78d3c5d064c6950f2dcb87e7e4d229af009b" exitCode=0 Dec 01 09:28:19 crc kubenswrapper[4873]: I1201 09:28:19.579050 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z85vq" event={"ID":"36b2e93f-f26a-4b05-83f9-3479f92ce038","Type":"ContainerDied","Data":"14fb93271d6a7e20e0456a629c4a78d3c5d064c6950f2dcb87e7e4d229af009b"} Dec 01 09:28:20 crc kubenswrapper[4873]: I1201 09:28:20.608153 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z85vq" event={"ID":"36b2e93f-f26a-4b05-83f9-3479f92ce038","Type":"ContainerStarted","Data":"ef68547a5f0564f4ebd8555c369cfa7f9132506cf6b066a9d6d8e9138f8eef38"} Dec 01 09:28:20 crc kubenswrapper[4873]: I1201 09:28:20.636281 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-z85vq" podStartSLOduration=2.844568334 podStartE2EDuration="5.63625145s" podCreationTimestamp="2025-12-01 09:28:15 +0000 UTC" firstStartedPulling="2025-12-01 09:28:17.558898511 +0000 UTC m=+2873.461007050" lastFinishedPulling="2025-12-01 09:28:20.350581637 +0000 UTC m=+2876.252690166" observedRunningTime="2025-12-01 09:28:20.630695051 +0000 UTC m=+2876.532803600" watchObservedRunningTime="2025-12-01 09:28:20.63625145 +0000 UTC m=+2876.538359989" Dec 01 09:28:25 crc kubenswrapper[4873]: I1201 09:28:25.759608 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-z85vq" Dec 01 09:28:25 crc kubenswrapper[4873]: I1201 09:28:25.760366 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-z85vq" Dec 01 09:28:25 crc kubenswrapper[4873]: I1201 09:28:25.810478 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-z85vq" Dec 01 09:28:26 crc kubenswrapper[4873]: I1201 09:28:26.724112 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-z85vq" Dec 01 09:28:26 crc kubenswrapper[4873]: I1201 09:28:26.785951 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-z85vq"] Dec 01 09:28:28 crc kubenswrapper[4873]: I1201 09:28:28.692086 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-z85vq" podUID="36b2e93f-f26a-4b05-83f9-3479f92ce038" containerName="registry-server" containerID="cri-o://ef68547a5f0564f4ebd8555c369cfa7f9132506cf6b066a9d6d8e9138f8eef38" gracePeriod=2 Dec 01 09:28:29 crc kubenswrapper[4873]: I1201 09:28:29.709320 4873 generic.go:334] "Generic (PLEG): container finished" podID="36b2e93f-f26a-4b05-83f9-3479f92ce038" containerID="ef68547a5f0564f4ebd8555c369cfa7f9132506cf6b066a9d6d8e9138f8eef38" exitCode=0 Dec 01 09:28:29 crc kubenswrapper[4873]: I1201 09:28:29.709416 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z85vq" event={"ID":"36b2e93f-f26a-4b05-83f9-3479f92ce038","Type":"ContainerDied","Data":"ef68547a5f0564f4ebd8555c369cfa7f9132506cf6b066a9d6d8e9138f8eef38"} Dec 01 09:28:29 crc kubenswrapper[4873]: I1201 09:28:29.709871 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z85vq" event={"ID":"36b2e93f-f26a-4b05-83f9-3479f92ce038","Type":"ContainerDied","Data":"0f503c68fd10a4ae393772931169bee60d6194931eaad22416870f38c0070b0d"} Dec 01 09:28:29 crc kubenswrapper[4873]: I1201 09:28:29.709899 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0f503c68fd10a4ae393772931169bee60d6194931eaad22416870f38c0070b0d" Dec 01 09:28:29 crc kubenswrapper[4873]: I1201 09:28:29.766392 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z85vq" Dec 01 09:28:29 crc kubenswrapper[4873]: I1201 09:28:29.880764 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36b2e93f-f26a-4b05-83f9-3479f92ce038-catalog-content\") pod \"36b2e93f-f26a-4b05-83f9-3479f92ce038\" (UID: \"36b2e93f-f26a-4b05-83f9-3479f92ce038\") " Dec 01 09:28:29 crc kubenswrapper[4873]: I1201 09:28:29.881653 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7kw9\" (UniqueName: \"kubernetes.io/projected/36b2e93f-f26a-4b05-83f9-3479f92ce038-kube-api-access-r7kw9\") pod \"36b2e93f-f26a-4b05-83f9-3479f92ce038\" (UID: \"36b2e93f-f26a-4b05-83f9-3479f92ce038\") " Dec 01 09:28:29 crc kubenswrapper[4873]: I1201 09:28:29.881821 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36b2e93f-f26a-4b05-83f9-3479f92ce038-utilities\") pod \"36b2e93f-f26a-4b05-83f9-3479f92ce038\" (UID: \"36b2e93f-f26a-4b05-83f9-3479f92ce038\") " Dec 01 09:28:29 crc kubenswrapper[4873]: I1201 09:28:29.883004 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36b2e93f-f26a-4b05-83f9-3479f92ce038-utilities" (OuterVolumeSpecName: "utilities") pod "36b2e93f-f26a-4b05-83f9-3479f92ce038" (UID: "36b2e93f-f26a-4b05-83f9-3479f92ce038"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:28:29 crc kubenswrapper[4873]: I1201 09:28:29.890263 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36b2e93f-f26a-4b05-83f9-3479f92ce038-kube-api-access-r7kw9" (OuterVolumeSpecName: "kube-api-access-r7kw9") pod "36b2e93f-f26a-4b05-83f9-3479f92ce038" (UID: "36b2e93f-f26a-4b05-83f9-3479f92ce038"). InnerVolumeSpecName "kube-api-access-r7kw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:28:29 crc kubenswrapper[4873]: I1201 09:28:29.939282 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36b2e93f-f26a-4b05-83f9-3479f92ce038-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "36b2e93f-f26a-4b05-83f9-3479f92ce038" (UID: "36b2e93f-f26a-4b05-83f9-3479f92ce038"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:28:29 crc kubenswrapper[4873]: I1201 09:28:29.985064 4873 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36b2e93f-f26a-4b05-83f9-3479f92ce038-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 09:28:29 crc kubenswrapper[4873]: I1201 09:28:29.985131 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7kw9\" (UniqueName: \"kubernetes.io/projected/36b2e93f-f26a-4b05-83f9-3479f92ce038-kube-api-access-r7kw9\") on node \"crc\" DevicePath \"\"" Dec 01 09:28:29 crc kubenswrapper[4873]: I1201 09:28:29.985146 4873 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36b2e93f-f26a-4b05-83f9-3479f92ce038-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 09:28:30 crc kubenswrapper[4873]: I1201 09:28:30.719995 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z85vq" Dec 01 09:28:30 crc kubenswrapper[4873]: I1201 09:28:30.748647 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-z85vq"] Dec 01 09:28:30 crc kubenswrapper[4873]: I1201 09:28:30.759669 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-z85vq"] Dec 01 09:28:31 crc kubenswrapper[4873]: I1201 09:28:31.059815 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:28:31 crc kubenswrapper[4873]: I1201 09:28:31.060264 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:28:31 crc kubenswrapper[4873]: I1201 09:28:31.060317 4873 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" Dec 01 09:28:31 crc kubenswrapper[4873]: I1201 09:28:31.061326 4873 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1eb6abed9ec75ad80889e13d075ac04aa13db51de45fa2f980efa98223444319"} pod="openshift-machine-config-operator/machine-config-daemon-scwpp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 09:28:31 crc kubenswrapper[4873]: I1201 09:28:31.061390 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" containerID="cri-o://1eb6abed9ec75ad80889e13d075ac04aa13db51de45fa2f980efa98223444319" gracePeriod=600 Dec 01 09:28:31 crc kubenswrapper[4873]: E1201 09:28:31.727411 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:28:31 crc kubenswrapper[4873]: I1201 09:28:31.753221 4873 generic.go:334] "Generic (PLEG): container finished" podID="fef7b114-0e07-402d-a37b-315c36011f4b" containerID="1eb6abed9ec75ad80889e13d075ac04aa13db51de45fa2f980efa98223444319" exitCode=0 Dec 01 09:28:31 crc kubenswrapper[4873]: I1201 09:28:31.753365 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" event={"ID":"fef7b114-0e07-402d-a37b-315c36011f4b","Type":"ContainerDied","Data":"1eb6abed9ec75ad80889e13d075ac04aa13db51de45fa2f980efa98223444319"} Dec 01 09:28:31 crc kubenswrapper[4873]: I1201 09:28:31.753437 4873 scope.go:117] "RemoveContainer" containerID="5086f7da0c61e3334bc2085e8599b53512aec9e3eb10321166ea8d6e4ebb4c9d" Dec 01 09:28:31 crc kubenswrapper[4873]: I1201 09:28:31.754890 4873 scope.go:117] "RemoveContainer" containerID="1eb6abed9ec75ad80889e13d075ac04aa13db51de45fa2f980efa98223444319" Dec 01 09:28:31 crc kubenswrapper[4873]: E1201 09:28:31.755483 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:28:32 crc kubenswrapper[4873]: I1201 09:28:32.441421 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36b2e93f-f26a-4b05-83f9-3479f92ce038" path="/var/lib/kubelet/pods/36b2e93f-f26a-4b05-83f9-3479f92ce038/volumes" Dec 01 09:28:43 crc kubenswrapper[4873]: I1201 09:28:43.430902 4873 scope.go:117] "RemoveContainer" containerID="1eb6abed9ec75ad80889e13d075ac04aa13db51de45fa2f980efa98223444319" Dec 01 09:28:43 crc kubenswrapper[4873]: E1201 09:28:43.432314 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:28:55 crc kubenswrapper[4873]: I1201 09:28:55.430266 4873 scope.go:117] "RemoveContainer" containerID="1eb6abed9ec75ad80889e13d075ac04aa13db51de45fa2f980efa98223444319" Dec 01 09:28:55 crc kubenswrapper[4873]: E1201 09:28:55.431215 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:28:59 crc kubenswrapper[4873]: I1201 09:28:59.425573 4873 scope.go:117] "RemoveContainer" containerID="0b93d4ae43167f894e68a1226cc5a2d30889d3b80794e27529f8ae3ac9284987" Dec 01 09:28:59 crc kubenswrapper[4873]: I1201 09:28:59.534696 4873 scope.go:117] "RemoveContainer" containerID="f165854cd4bc8b7a60aa520b4bdf503536cc30f8ac115ce50673b958ddd9c771" Dec 01 09:28:59 crc kubenswrapper[4873]: I1201 09:28:59.573090 4873 scope.go:117] "RemoveContainer" containerID="6290fe887ebefaca070523485158812ab8b791bb81e722a43bc8de21c2f184e6" Dec 01 09:29:08 crc kubenswrapper[4873]: I1201 09:29:08.431107 4873 scope.go:117] "RemoveContainer" containerID="1eb6abed9ec75ad80889e13d075ac04aa13db51de45fa2f980efa98223444319" Dec 01 09:29:08 crc kubenswrapper[4873]: E1201 09:29:08.433406 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:29:19 crc kubenswrapper[4873]: I1201 09:29:19.431469 4873 scope.go:117] "RemoveContainer" containerID="1eb6abed9ec75ad80889e13d075ac04aa13db51de45fa2f980efa98223444319" Dec 01 09:29:19 crc kubenswrapper[4873]: E1201 09:29:19.433193 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:29:24 crc kubenswrapper[4873]: I1201 09:29:24.796516 4873 generic.go:334] "Generic (PLEG): container finished" podID="328559dd-d6ca-400a-9bb1-955781e2e1ea" containerID="c834e261924bb4cd420b71da1622d43fbc94e7692cbe694dc457544c4c9b46bb" exitCode=0 Dec 01 09:29:24 crc kubenswrapper[4873]: I1201 09:29:24.797112 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp" event={"ID":"328559dd-d6ca-400a-9bb1-955781e2e1ea","Type":"ContainerDied","Data":"c834e261924bb4cd420b71da1622d43fbc94e7692cbe694dc457544c4c9b46bb"} Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.250377 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp" Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.357746 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/328559dd-d6ca-400a-9bb1-955781e2e1ea-ceph\") pod \"328559dd-d6ca-400a-9bb1-955781e2e1ea\" (UID: \"328559dd-d6ca-400a-9bb1-955781e2e1ea\") " Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.357814 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/328559dd-d6ca-400a-9bb1-955781e2e1ea-ssh-key\") pod \"328559dd-d6ca-400a-9bb1-955781e2e1ea\" (UID: \"328559dd-d6ca-400a-9bb1-955781e2e1ea\") " Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.358099 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/328559dd-d6ca-400a-9bb1-955781e2e1ea-libvirt-secret-0\") pod \"328559dd-d6ca-400a-9bb1-955781e2e1ea\" (UID: \"328559dd-d6ca-400a-9bb1-955781e2e1ea\") " Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.358155 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/328559dd-d6ca-400a-9bb1-955781e2e1ea-libvirt-combined-ca-bundle\") pod \"328559dd-d6ca-400a-9bb1-955781e2e1ea\" (UID: \"328559dd-d6ca-400a-9bb1-955781e2e1ea\") " Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.358210 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/328559dd-d6ca-400a-9bb1-955781e2e1ea-inventory\") pod \"328559dd-d6ca-400a-9bb1-955781e2e1ea\" (UID: \"328559dd-d6ca-400a-9bb1-955781e2e1ea\") " Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.358309 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qdcxs\" (UniqueName: \"kubernetes.io/projected/328559dd-d6ca-400a-9bb1-955781e2e1ea-kube-api-access-qdcxs\") pod \"328559dd-d6ca-400a-9bb1-955781e2e1ea\" (UID: \"328559dd-d6ca-400a-9bb1-955781e2e1ea\") " Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.366728 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/328559dd-d6ca-400a-9bb1-955781e2e1ea-kube-api-access-qdcxs" (OuterVolumeSpecName: "kube-api-access-qdcxs") pod "328559dd-d6ca-400a-9bb1-955781e2e1ea" (UID: "328559dd-d6ca-400a-9bb1-955781e2e1ea"). InnerVolumeSpecName "kube-api-access-qdcxs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.369279 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/328559dd-d6ca-400a-9bb1-955781e2e1ea-ceph" (OuterVolumeSpecName: "ceph") pod "328559dd-d6ca-400a-9bb1-955781e2e1ea" (UID: "328559dd-d6ca-400a-9bb1-955781e2e1ea"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.375544 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/328559dd-d6ca-400a-9bb1-955781e2e1ea-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "328559dd-d6ca-400a-9bb1-955781e2e1ea" (UID: "328559dd-d6ca-400a-9bb1-955781e2e1ea"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.391897 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/328559dd-d6ca-400a-9bb1-955781e2e1ea-inventory" (OuterVolumeSpecName: "inventory") pod "328559dd-d6ca-400a-9bb1-955781e2e1ea" (UID: "328559dd-d6ca-400a-9bb1-955781e2e1ea"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.391957 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/328559dd-d6ca-400a-9bb1-955781e2e1ea-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "328559dd-d6ca-400a-9bb1-955781e2e1ea" (UID: "328559dd-d6ca-400a-9bb1-955781e2e1ea"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.392855 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/328559dd-d6ca-400a-9bb1-955781e2e1ea-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "328559dd-d6ca-400a-9bb1-955781e2e1ea" (UID: "328559dd-d6ca-400a-9bb1-955781e2e1ea"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.461151 4873 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/328559dd-d6ca-400a-9bb1-955781e2e1ea-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.461301 4873 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/328559dd-d6ca-400a-9bb1-955781e2e1ea-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.461382 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qdcxs\" (UniqueName: \"kubernetes.io/projected/328559dd-d6ca-400a-9bb1-955781e2e1ea-kube-api-access-qdcxs\") on node \"crc\" DevicePath \"\"" Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.461452 4873 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/328559dd-d6ca-400a-9bb1-955781e2e1ea-ceph\") on node \"crc\" DevicePath \"\"" Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.461510 4873 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/328559dd-d6ca-400a-9bb1-955781e2e1ea-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.461589 4873 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/328559dd-d6ca-400a-9bb1-955781e2e1ea-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.827223 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp" event={"ID":"328559dd-d6ca-400a-9bb1-955781e2e1ea","Type":"ContainerDied","Data":"a88250fe25d343a65ff9649e1baedf912e4e13be7f730b6d1a0c8992e987e074"} Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.827282 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a88250fe25d343a65ff9649e1baedf912e4e13be7f730b6d1a0c8992e987e074" Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.827279 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp" Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.932675 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp"] Dec 01 09:29:26 crc kubenswrapper[4873]: E1201 09:29:26.934070 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36b2e93f-f26a-4b05-83f9-3479f92ce038" containerName="extract-utilities" Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.934129 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="36b2e93f-f26a-4b05-83f9-3479f92ce038" containerName="extract-utilities" Dec 01 09:29:26 crc kubenswrapper[4873]: E1201 09:29:26.934144 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36b2e93f-f26a-4b05-83f9-3479f92ce038" containerName="registry-server" Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.934152 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="36b2e93f-f26a-4b05-83f9-3479f92ce038" containerName="registry-server" Dec 01 09:29:26 crc kubenswrapper[4873]: E1201 09:29:26.934180 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="328559dd-d6ca-400a-9bb1-955781e2e1ea" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.934191 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="328559dd-d6ca-400a-9bb1-955781e2e1ea" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 01 09:29:26 crc kubenswrapper[4873]: E1201 09:29:26.934206 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36b2e93f-f26a-4b05-83f9-3479f92ce038" containerName="extract-content" Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.934217 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="36b2e93f-f26a-4b05-83f9-3479f92ce038" containerName="extract-content" Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.934433 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="36b2e93f-f26a-4b05-83f9-3479f92ce038" containerName="registry-server" Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.934449 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="328559dd-d6ca-400a-9bb1-955781e2e1ea" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.935565 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp" Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.938826 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.939120 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ceph-nova" Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.939221 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.939500 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.941861 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.942614 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9ghqw" Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.942736 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.942899 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.943267 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.945695 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp"] Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.970858 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp\" (UID: \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp" Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.971302 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp\" (UID: \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp" Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.971556 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp\" (UID: \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp" Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.971687 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h55x4\" (UniqueName: \"kubernetes.io/projected/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-kube-api-access-h55x4\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp\" (UID: \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp" Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.971867 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp\" (UID: \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp" Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.971994 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp\" (UID: \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp" Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.972137 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp\" (UID: \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp" Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.972261 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp\" (UID: \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp" Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.972437 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp\" (UID: \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp" Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.972572 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp\" (UID: \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp" Dec 01 09:29:26 crc kubenswrapper[4873]: I1201 09:29:26.972884 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp\" (UID: \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp" Dec 01 09:29:27 crc kubenswrapper[4873]: I1201 09:29:27.075726 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp\" (UID: \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp" Dec 01 09:29:27 crc kubenswrapper[4873]: I1201 09:29:27.075813 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h55x4\" (UniqueName: \"kubernetes.io/projected/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-kube-api-access-h55x4\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp\" (UID: \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp" Dec 01 09:29:27 crc kubenswrapper[4873]: I1201 09:29:27.075878 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp\" (UID: \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp" Dec 01 09:29:27 crc kubenswrapper[4873]: I1201 09:29:27.075917 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp\" (UID: \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp" Dec 01 09:29:27 crc kubenswrapper[4873]: I1201 09:29:27.075944 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp\" (UID: \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp" Dec 01 09:29:27 crc kubenswrapper[4873]: I1201 09:29:27.075971 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp\" (UID: \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp" Dec 01 09:29:27 crc kubenswrapper[4873]: I1201 09:29:27.076060 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp\" (UID: \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp" Dec 01 09:29:27 crc kubenswrapper[4873]: I1201 09:29:27.076094 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp\" (UID: \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp" Dec 01 09:29:27 crc kubenswrapper[4873]: I1201 09:29:27.076120 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp\" (UID: \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp" Dec 01 09:29:27 crc kubenswrapper[4873]: I1201 09:29:27.076159 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp\" (UID: \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp" Dec 01 09:29:27 crc kubenswrapper[4873]: I1201 09:29:27.076192 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp\" (UID: \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp" Dec 01 09:29:27 crc kubenswrapper[4873]: I1201 09:29:27.081850 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp\" (UID: \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp" Dec 01 09:29:27 crc kubenswrapper[4873]: I1201 09:29:27.082717 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp\" (UID: \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp" Dec 01 09:29:27 crc kubenswrapper[4873]: I1201 09:29:27.083198 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp\" (UID: \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp" Dec 01 09:29:27 crc kubenswrapper[4873]: I1201 09:29:27.086158 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp\" (UID: \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp" Dec 01 09:29:27 crc kubenswrapper[4873]: I1201 09:29:27.086670 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp\" (UID: \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp" Dec 01 09:29:27 crc kubenswrapper[4873]: I1201 09:29:27.086832 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp\" (UID: \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp" Dec 01 09:29:27 crc kubenswrapper[4873]: I1201 09:29:27.088742 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp\" (UID: \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp" Dec 01 09:29:27 crc kubenswrapper[4873]: I1201 09:29:27.089306 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp\" (UID: \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp" Dec 01 09:29:27 crc kubenswrapper[4873]: I1201 09:29:27.089912 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp\" (UID: \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp" Dec 01 09:29:27 crc kubenswrapper[4873]: I1201 09:29:27.099127 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h55x4\" (UniqueName: \"kubernetes.io/projected/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-kube-api-access-h55x4\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp\" (UID: \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp" Dec 01 09:29:27 crc kubenswrapper[4873]: I1201 09:29:27.117751 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp\" (UID: \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp" Dec 01 09:29:27 crc kubenswrapper[4873]: I1201 09:29:27.255445 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp" Dec 01 09:29:27 crc kubenswrapper[4873]: I1201 09:29:27.895797 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp"] Dec 01 09:29:28 crc kubenswrapper[4873]: I1201 09:29:28.851379 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp" event={"ID":"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf","Type":"ContainerStarted","Data":"d8471494e44ba59c9f556beb03116bbfd8b68e097210ca55e5090b2226c66462"} Dec 01 09:29:28 crc kubenswrapper[4873]: I1201 09:29:28.851826 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp" event={"ID":"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf","Type":"ContainerStarted","Data":"68ecc9f2ffe8058c78b8df2aef71f51a3cd2d8a387900d424b2fef98ac76e79d"} Dec 01 09:29:28 crc kubenswrapper[4873]: I1201 09:29:28.871227 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp" podStartSLOduration=2.303490948 podStartE2EDuration="2.871205944s" podCreationTimestamp="2025-12-01 09:29:26 +0000 UTC" firstStartedPulling="2025-12-01 09:29:27.918006759 +0000 UTC m=+2943.820115298" lastFinishedPulling="2025-12-01 09:29:28.485721755 +0000 UTC m=+2944.387830294" observedRunningTime="2025-12-01 09:29:28.870659501 +0000 UTC m=+2944.772768050" watchObservedRunningTime="2025-12-01 09:29:28.871205944 +0000 UTC m=+2944.773314483" Dec 01 09:29:31 crc kubenswrapper[4873]: I1201 09:29:31.431044 4873 scope.go:117] "RemoveContainer" containerID="1eb6abed9ec75ad80889e13d075ac04aa13db51de45fa2f980efa98223444319" Dec 01 09:29:31 crc kubenswrapper[4873]: E1201 09:29:31.431753 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:29:44 crc kubenswrapper[4873]: I1201 09:29:44.438124 4873 scope.go:117] "RemoveContainer" containerID="1eb6abed9ec75ad80889e13d075ac04aa13db51de45fa2f980efa98223444319" Dec 01 09:29:44 crc kubenswrapper[4873]: E1201 09:29:44.438956 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:29:59 crc kubenswrapper[4873]: I1201 09:29:59.429918 4873 scope.go:117] "RemoveContainer" containerID="1eb6abed9ec75ad80889e13d075ac04aa13db51de45fa2f980efa98223444319" Dec 01 09:29:59 crc kubenswrapper[4873]: E1201 09:29:59.430992 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:30:00 crc kubenswrapper[4873]: I1201 09:30:00.160638 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409690-wfrsd"] Dec 01 09:30:00 crc kubenswrapper[4873]: I1201 09:30:00.162640 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409690-wfrsd" Dec 01 09:30:00 crc kubenswrapper[4873]: I1201 09:30:00.173728 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 09:30:00 crc kubenswrapper[4873]: I1201 09:30:00.173913 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 09:30:00 crc kubenswrapper[4873]: I1201 09:30:00.177791 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409690-wfrsd"] Dec 01 09:30:00 crc kubenswrapper[4873]: I1201 09:30:00.257456 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grdmf\" (UniqueName: \"kubernetes.io/projected/3e8e7d3b-9356-4841-9014-3a86d8a77540-kube-api-access-grdmf\") pod \"collect-profiles-29409690-wfrsd\" (UID: \"3e8e7d3b-9356-4841-9014-3a86d8a77540\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409690-wfrsd" Dec 01 09:30:00 crc kubenswrapper[4873]: I1201 09:30:00.257533 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3e8e7d3b-9356-4841-9014-3a86d8a77540-config-volume\") pod \"collect-profiles-29409690-wfrsd\" (UID: \"3e8e7d3b-9356-4841-9014-3a86d8a77540\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409690-wfrsd" Dec 01 09:30:00 crc kubenswrapper[4873]: I1201 09:30:00.257783 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3e8e7d3b-9356-4841-9014-3a86d8a77540-secret-volume\") pod \"collect-profiles-29409690-wfrsd\" (UID: \"3e8e7d3b-9356-4841-9014-3a86d8a77540\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409690-wfrsd" Dec 01 09:30:00 crc kubenswrapper[4873]: I1201 09:30:00.360453 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3e8e7d3b-9356-4841-9014-3a86d8a77540-config-volume\") pod \"collect-profiles-29409690-wfrsd\" (UID: \"3e8e7d3b-9356-4841-9014-3a86d8a77540\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409690-wfrsd" Dec 01 09:30:00 crc kubenswrapper[4873]: I1201 09:30:00.360606 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3e8e7d3b-9356-4841-9014-3a86d8a77540-secret-volume\") pod \"collect-profiles-29409690-wfrsd\" (UID: \"3e8e7d3b-9356-4841-9014-3a86d8a77540\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409690-wfrsd" Dec 01 09:30:00 crc kubenswrapper[4873]: I1201 09:30:00.360725 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grdmf\" (UniqueName: \"kubernetes.io/projected/3e8e7d3b-9356-4841-9014-3a86d8a77540-kube-api-access-grdmf\") pod \"collect-profiles-29409690-wfrsd\" (UID: \"3e8e7d3b-9356-4841-9014-3a86d8a77540\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409690-wfrsd" Dec 01 09:30:00 crc kubenswrapper[4873]: I1201 09:30:00.361692 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3e8e7d3b-9356-4841-9014-3a86d8a77540-config-volume\") pod \"collect-profiles-29409690-wfrsd\" (UID: \"3e8e7d3b-9356-4841-9014-3a86d8a77540\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409690-wfrsd" Dec 01 09:30:00 crc kubenswrapper[4873]: I1201 09:30:00.369116 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3e8e7d3b-9356-4841-9014-3a86d8a77540-secret-volume\") pod \"collect-profiles-29409690-wfrsd\" (UID: \"3e8e7d3b-9356-4841-9014-3a86d8a77540\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409690-wfrsd" Dec 01 09:30:00 crc kubenswrapper[4873]: I1201 09:30:00.385517 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grdmf\" (UniqueName: \"kubernetes.io/projected/3e8e7d3b-9356-4841-9014-3a86d8a77540-kube-api-access-grdmf\") pod \"collect-profiles-29409690-wfrsd\" (UID: \"3e8e7d3b-9356-4841-9014-3a86d8a77540\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409690-wfrsd" Dec 01 09:30:00 crc kubenswrapper[4873]: I1201 09:30:00.511656 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409690-wfrsd" Dec 01 09:30:01 crc kubenswrapper[4873]: I1201 09:30:00.999619 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409690-wfrsd"] Dec 01 09:30:01 crc kubenswrapper[4873]: I1201 09:30:01.193896 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409690-wfrsd" event={"ID":"3e8e7d3b-9356-4841-9014-3a86d8a77540","Type":"ContainerStarted","Data":"d0bfe1270f9d5ad97d6789abb39fd3e484932f8e3d8c033966408153cba0fee6"} Dec 01 09:30:01 crc kubenswrapper[4873]: I1201 09:30:01.218404 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29409690-wfrsd" podStartSLOduration=1.218376705 podStartE2EDuration="1.218376705s" podCreationTimestamp="2025-12-01 09:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:30:01.216860617 +0000 UTC m=+2977.118969166" watchObservedRunningTime="2025-12-01 09:30:01.218376705 +0000 UTC m=+2977.120485244" Dec 01 09:30:02 crc kubenswrapper[4873]: I1201 09:30:02.206638 4873 generic.go:334] "Generic (PLEG): container finished" podID="3e8e7d3b-9356-4841-9014-3a86d8a77540" containerID="3e60da5faca15daee01020f7988a194ec6d53b79ab3ace8144c1607307f8e038" exitCode=0 Dec 01 09:30:02 crc kubenswrapper[4873]: I1201 09:30:02.206739 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409690-wfrsd" event={"ID":"3e8e7d3b-9356-4841-9014-3a86d8a77540","Type":"ContainerDied","Data":"3e60da5faca15daee01020f7988a194ec6d53b79ab3ace8144c1607307f8e038"} Dec 01 09:30:03 crc kubenswrapper[4873]: I1201 09:30:03.557532 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409690-wfrsd" Dec 01 09:30:03 crc kubenswrapper[4873]: I1201 09:30:03.628925 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grdmf\" (UniqueName: \"kubernetes.io/projected/3e8e7d3b-9356-4841-9014-3a86d8a77540-kube-api-access-grdmf\") pod \"3e8e7d3b-9356-4841-9014-3a86d8a77540\" (UID: \"3e8e7d3b-9356-4841-9014-3a86d8a77540\") " Dec 01 09:30:03 crc kubenswrapper[4873]: I1201 09:30:03.629115 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3e8e7d3b-9356-4841-9014-3a86d8a77540-config-volume\") pod \"3e8e7d3b-9356-4841-9014-3a86d8a77540\" (UID: \"3e8e7d3b-9356-4841-9014-3a86d8a77540\") " Dec 01 09:30:03 crc kubenswrapper[4873]: I1201 09:30:03.629230 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3e8e7d3b-9356-4841-9014-3a86d8a77540-secret-volume\") pod \"3e8e7d3b-9356-4841-9014-3a86d8a77540\" (UID: \"3e8e7d3b-9356-4841-9014-3a86d8a77540\") " Dec 01 09:30:03 crc kubenswrapper[4873]: I1201 09:30:03.630453 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e8e7d3b-9356-4841-9014-3a86d8a77540-config-volume" (OuterVolumeSpecName: "config-volume") pod "3e8e7d3b-9356-4841-9014-3a86d8a77540" (UID: "3e8e7d3b-9356-4841-9014-3a86d8a77540"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:30:03 crc kubenswrapper[4873]: I1201 09:30:03.637703 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e8e7d3b-9356-4841-9014-3a86d8a77540-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3e8e7d3b-9356-4841-9014-3a86d8a77540" (UID: "3e8e7d3b-9356-4841-9014-3a86d8a77540"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:30:03 crc kubenswrapper[4873]: I1201 09:30:03.640766 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e8e7d3b-9356-4841-9014-3a86d8a77540-kube-api-access-grdmf" (OuterVolumeSpecName: "kube-api-access-grdmf") pod "3e8e7d3b-9356-4841-9014-3a86d8a77540" (UID: "3e8e7d3b-9356-4841-9014-3a86d8a77540"). InnerVolumeSpecName "kube-api-access-grdmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:30:03 crc kubenswrapper[4873]: I1201 09:30:03.731185 4873 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3e8e7d3b-9356-4841-9014-3a86d8a77540-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 09:30:03 crc kubenswrapper[4873]: I1201 09:30:03.731230 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grdmf\" (UniqueName: \"kubernetes.io/projected/3e8e7d3b-9356-4841-9014-3a86d8a77540-kube-api-access-grdmf\") on node \"crc\" DevicePath \"\"" Dec 01 09:30:03 crc kubenswrapper[4873]: I1201 09:30:03.731240 4873 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3e8e7d3b-9356-4841-9014-3a86d8a77540-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 09:30:04 crc kubenswrapper[4873]: I1201 09:30:04.230060 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409690-wfrsd" event={"ID":"3e8e7d3b-9356-4841-9014-3a86d8a77540","Type":"ContainerDied","Data":"d0bfe1270f9d5ad97d6789abb39fd3e484932f8e3d8c033966408153cba0fee6"} Dec 01 09:30:04 crc kubenswrapper[4873]: I1201 09:30:04.230118 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d0bfe1270f9d5ad97d6789abb39fd3e484932f8e3d8c033966408153cba0fee6" Dec 01 09:30:04 crc kubenswrapper[4873]: I1201 09:30:04.230163 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409690-wfrsd" Dec 01 09:30:04 crc kubenswrapper[4873]: I1201 09:30:04.305775 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409645-vfsz2"] Dec 01 09:30:04 crc kubenswrapper[4873]: I1201 09:30:04.315854 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409645-vfsz2"] Dec 01 09:30:04 crc kubenswrapper[4873]: I1201 09:30:04.442755 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6d9174b-c4b0-4176-9e95-291c04608323" path="/var/lib/kubelet/pods/f6d9174b-c4b0-4176-9e95-291c04608323/volumes" Dec 01 09:30:14 crc kubenswrapper[4873]: I1201 09:30:14.439118 4873 scope.go:117] "RemoveContainer" containerID="1eb6abed9ec75ad80889e13d075ac04aa13db51de45fa2f980efa98223444319" Dec 01 09:30:14 crc kubenswrapper[4873]: E1201 09:30:14.440609 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:30:28 crc kubenswrapper[4873]: I1201 09:30:28.430897 4873 scope.go:117] "RemoveContainer" containerID="1eb6abed9ec75ad80889e13d075ac04aa13db51de45fa2f980efa98223444319" Dec 01 09:30:28 crc kubenswrapper[4873]: E1201 09:30:28.432313 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:30:43 crc kubenswrapper[4873]: I1201 09:30:43.430723 4873 scope.go:117] "RemoveContainer" containerID="1eb6abed9ec75ad80889e13d075ac04aa13db51de45fa2f980efa98223444319" Dec 01 09:30:43 crc kubenswrapper[4873]: E1201 09:30:43.432340 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:30:54 crc kubenswrapper[4873]: I1201 09:30:54.437683 4873 scope.go:117] "RemoveContainer" containerID="1eb6abed9ec75ad80889e13d075ac04aa13db51de45fa2f980efa98223444319" Dec 01 09:30:54 crc kubenswrapper[4873]: E1201 09:30:54.440327 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:30:59 crc kubenswrapper[4873]: I1201 09:30:59.658977 4873 scope.go:117] "RemoveContainer" containerID="dd3d81fd19b458c0b28f082da23039ebaaab818011a7afc53a450012b4b8c01b" Dec 01 09:31:06 crc kubenswrapper[4873]: I1201 09:31:06.430594 4873 scope.go:117] "RemoveContainer" containerID="1eb6abed9ec75ad80889e13d075ac04aa13db51de45fa2f980efa98223444319" Dec 01 09:31:06 crc kubenswrapper[4873]: E1201 09:31:06.431785 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:31:21 crc kubenswrapper[4873]: I1201 09:31:21.430538 4873 scope.go:117] "RemoveContainer" containerID="1eb6abed9ec75ad80889e13d075ac04aa13db51de45fa2f980efa98223444319" Dec 01 09:31:21 crc kubenswrapper[4873]: E1201 09:31:21.431778 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:31:34 crc kubenswrapper[4873]: I1201 09:31:34.438157 4873 scope.go:117] "RemoveContainer" containerID="1eb6abed9ec75ad80889e13d075ac04aa13db51de45fa2f980efa98223444319" Dec 01 09:31:34 crc kubenswrapper[4873]: E1201 09:31:34.439255 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:31:48 crc kubenswrapper[4873]: I1201 09:31:48.430215 4873 scope.go:117] "RemoveContainer" containerID="1eb6abed9ec75ad80889e13d075ac04aa13db51de45fa2f980efa98223444319" Dec 01 09:31:48 crc kubenswrapper[4873]: E1201 09:31:48.431066 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:32:03 crc kubenswrapper[4873]: I1201 09:32:03.430228 4873 scope.go:117] "RemoveContainer" containerID="1eb6abed9ec75ad80889e13d075ac04aa13db51de45fa2f980efa98223444319" Dec 01 09:32:03 crc kubenswrapper[4873]: E1201 09:32:03.431118 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:32:15 crc kubenswrapper[4873]: I1201 09:32:15.430065 4873 scope.go:117] "RemoveContainer" containerID="1eb6abed9ec75ad80889e13d075ac04aa13db51de45fa2f980efa98223444319" Dec 01 09:32:15 crc kubenswrapper[4873]: E1201 09:32:15.430941 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:32:27 crc kubenswrapper[4873]: I1201 09:32:27.787793 4873 generic.go:334] "Generic (PLEG): container finished" podID="38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf" containerID="d8471494e44ba59c9f556beb03116bbfd8b68e097210ca55e5090b2226c66462" exitCode=0 Dec 01 09:32:27 crc kubenswrapper[4873]: I1201 09:32:27.787883 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp" event={"ID":"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf","Type":"ContainerDied","Data":"d8471494e44ba59c9f556beb03116bbfd8b68e097210ca55e5090b2226c66462"} Dec 01 09:32:29 crc kubenswrapper[4873]: I1201 09:32:29.310108 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp" Dec 01 09:32:29 crc kubenswrapper[4873]: I1201 09:32:29.431955 4873 scope.go:117] "RemoveContainer" containerID="1eb6abed9ec75ad80889e13d075ac04aa13db51de45fa2f980efa98223444319" Dec 01 09:32:29 crc kubenswrapper[4873]: E1201 09:32:29.433004 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:32:29 crc kubenswrapper[4873]: I1201 09:32:29.510223 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-ceph\") pod \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\" (UID: \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\") " Dec 01 09:32:29 crc kubenswrapper[4873]: I1201 09:32:29.510298 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-nova-cell1-compute-config-0\") pod \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\" (UID: \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\") " Dec 01 09:32:29 crc kubenswrapper[4873]: I1201 09:32:29.510380 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-nova-migration-ssh-key-1\") pod \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\" (UID: \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\") " Dec 01 09:32:29 crc kubenswrapper[4873]: I1201 09:32:29.510402 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-nova-migration-ssh-key-0\") pod \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\" (UID: \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\") " Dec 01 09:32:29 crc kubenswrapper[4873]: I1201 09:32:29.510440 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-nova-extra-config-0\") pod \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\" (UID: \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\") " Dec 01 09:32:29 crc kubenswrapper[4873]: I1201 09:32:29.511975 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-inventory\") pod \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\" (UID: \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\") " Dec 01 09:32:29 crc kubenswrapper[4873]: I1201 09:32:29.512149 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-nova-custom-ceph-combined-ca-bundle\") pod \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\" (UID: \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\") " Dec 01 09:32:29 crc kubenswrapper[4873]: I1201 09:32:29.512212 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h55x4\" (UniqueName: \"kubernetes.io/projected/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-kube-api-access-h55x4\") pod \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\" (UID: \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\") " Dec 01 09:32:29 crc kubenswrapper[4873]: I1201 09:32:29.512240 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-nova-cell1-compute-config-1\") pod \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\" (UID: \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\") " Dec 01 09:32:29 crc kubenswrapper[4873]: I1201 09:32:29.512306 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-ceph-nova-0\") pod \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\" (UID: \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\") " Dec 01 09:32:29 crc kubenswrapper[4873]: I1201 09:32:29.512336 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-ssh-key\") pod \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\" (UID: \"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf\") " Dec 01 09:32:29 crc kubenswrapper[4873]: I1201 09:32:29.518724 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-ceph" (OuterVolumeSpecName: "ceph") pod "38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf" (UID: "38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:32:29 crc kubenswrapper[4873]: I1201 09:32:29.524371 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-kube-api-access-h55x4" (OuterVolumeSpecName: "kube-api-access-h55x4") pod "38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf" (UID: "38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf"). InnerVolumeSpecName "kube-api-access-h55x4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:32:29 crc kubenswrapper[4873]: I1201 09:32:29.524380 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-nova-custom-ceph-combined-ca-bundle" (OuterVolumeSpecName: "nova-custom-ceph-combined-ca-bundle") pod "38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf" (UID: "38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf"). InnerVolumeSpecName "nova-custom-ceph-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:32:29 crc kubenswrapper[4873]: I1201 09:32:29.549794 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-inventory" (OuterVolumeSpecName: "inventory") pod "38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf" (UID: "38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:32:29 crc kubenswrapper[4873]: I1201 09:32:29.557102 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf" (UID: "38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:32:29 crc kubenswrapper[4873]: I1201 09:32:29.558473 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf" (UID: "38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:32:29 crc kubenswrapper[4873]: I1201 09:32:29.568345 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf" (UID: "38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:32:29 crc kubenswrapper[4873]: I1201 09:32:29.572187 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf" (UID: "38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:32:29 crc kubenswrapper[4873]: I1201 09:32:29.574390 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf" (UID: "38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:32:29 crc kubenswrapper[4873]: I1201 09:32:29.578908 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf" (UID: "38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:32:29 crc kubenswrapper[4873]: I1201 09:32:29.595092 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-ceph-nova-0" (OuterVolumeSpecName: "ceph-nova-0") pod "38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf" (UID: "38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf"). InnerVolumeSpecName "ceph-nova-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:32:29 crc kubenswrapper[4873]: I1201 09:32:29.617617 4873 reconciler_common.go:293] "Volume detached for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-nova-custom-ceph-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:29 crc kubenswrapper[4873]: I1201 09:32:29.617656 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h55x4\" (UniqueName: \"kubernetes.io/projected/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-kube-api-access-h55x4\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:29 crc kubenswrapper[4873]: I1201 09:32:29.617675 4873 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:29 crc kubenswrapper[4873]: I1201 09:32:29.617688 4873 reconciler_common.go:293] "Volume detached for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-ceph-nova-0\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:29 crc kubenswrapper[4873]: I1201 09:32:29.617700 4873 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:29 crc kubenswrapper[4873]: I1201 09:32:29.617711 4873 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-ceph\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:29 crc kubenswrapper[4873]: I1201 09:32:29.617722 4873 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:29 crc kubenswrapper[4873]: I1201 09:32:29.617734 4873 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:29 crc kubenswrapper[4873]: I1201 09:32:29.617747 4873 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:29 crc kubenswrapper[4873]: I1201 09:32:29.617760 4873 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:29 crc kubenswrapper[4873]: I1201 09:32:29.617774 4873 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:29 crc kubenswrapper[4873]: I1201 09:32:29.810211 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp" event={"ID":"38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf","Type":"ContainerDied","Data":"68ecc9f2ffe8058c78b8df2aef71f51a3cd2d8a387900d424b2fef98ac76e79d"} Dec 01 09:32:29 crc kubenswrapper[4873]: I1201 09:32:29.810269 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="68ecc9f2ffe8058c78b8df2aef71f51a3cd2d8a387900d424b2fef98ac76e79d" Dec 01 09:32:29 crc kubenswrapper[4873]: I1201 09:32:29.810291 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp" Dec 01 09:32:44 crc kubenswrapper[4873]: I1201 09:32:44.437377 4873 scope.go:117] "RemoveContainer" containerID="1eb6abed9ec75ad80889e13d075ac04aa13db51de45fa2f980efa98223444319" Dec 01 09:32:44 crc kubenswrapper[4873]: E1201 09:32:44.438563 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.411137 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-volume1-0"] Dec 01 09:32:45 crc kubenswrapper[4873]: E1201 09:32:45.411755 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e8e7d3b-9356-4841-9014-3a86d8a77540" containerName="collect-profiles" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.411780 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e8e7d3b-9356-4841-9014-3a86d8a77540" containerName="collect-profiles" Dec 01 09:32:45 crc kubenswrapper[4873]: E1201 09:32:45.411826 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.411841 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.412107 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e8e7d3b-9356-4841-9014-3a86d8a77540" containerName="collect-profiles" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.412136 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.413547 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.416857 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.420706 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-volume1-config-data" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.426382 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.466435 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.468637 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.472832 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.495908 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/96e55f05-de52-4914-9fc8-7634bff464a7-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.496007 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.496092 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-sys\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.496126 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/96e55f05-de52-4914-9fc8-7634bff464a7-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.496149 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/96e55f05-de52-4914-9fc8-7634bff464a7-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.496184 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96e55f05-de52-4914-9fc8-7634bff464a7-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.496218 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-lib-modules\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.496243 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-run\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.496276 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pg4hj\" (UniqueName: \"kubernetes.io/projected/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-kube-api-access-pg4hj\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.496350 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96e55f05-de52-4914-9fc8-7634bff464a7-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.496393 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/96e55f05-de52-4914-9fc8-7634bff464a7-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.496456 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-ceph\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.496526 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-scripts\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.496561 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/96e55f05-de52-4914-9fc8-7634bff464a7-run\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.496581 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-config-data-custom\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.496616 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwnhq\" (UniqueName: \"kubernetes.io/projected/96e55f05-de52-4914-9fc8-7634bff464a7-kube-api-access-kwnhq\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.496649 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/96e55f05-de52-4914-9fc8-7634bff464a7-dev\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.496669 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96e55f05-de52-4914-9fc8-7634bff464a7-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.496691 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.496715 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/96e55f05-de52-4914-9fc8-7634bff464a7-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.496804 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/96e55f05-de52-4914-9fc8-7634bff464a7-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.496917 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-config-data\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.496983 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-etc-nvme\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.497004 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.497043 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.497076 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/96e55f05-de52-4914-9fc8-7634bff464a7-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.497097 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-dev\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.497127 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/96e55f05-de52-4914-9fc8-7634bff464a7-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.497158 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/96e55f05-de52-4914-9fc8-7634bff464a7-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.497209 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.497229 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.497318 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/96e55f05-de52-4914-9fc8-7634bff464a7-sys\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.538266 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.599848 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-scripts\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.600202 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/96e55f05-de52-4914-9fc8-7634bff464a7-run\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.600450 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/96e55f05-de52-4914-9fc8-7634bff464a7-run\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.601453 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-config-data-custom\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.601618 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwnhq\" (UniqueName: \"kubernetes.io/projected/96e55f05-de52-4914-9fc8-7634bff464a7-kube-api-access-kwnhq\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.601665 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/96e55f05-de52-4914-9fc8-7634bff464a7-dev\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.601695 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96e55f05-de52-4914-9fc8-7634bff464a7-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.601716 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.601744 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/96e55f05-de52-4914-9fc8-7634bff464a7-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.601772 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/96e55f05-de52-4914-9fc8-7634bff464a7-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.601830 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-config-data\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.601884 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-etc-nvme\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.601908 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.601932 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.601975 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/96e55f05-de52-4914-9fc8-7634bff464a7-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.601994 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-dev\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.602045 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/96e55f05-de52-4914-9fc8-7634bff464a7-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.602076 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/96e55f05-de52-4914-9fc8-7634bff464a7-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.602135 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.602160 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.602296 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/96e55f05-de52-4914-9fc8-7634bff464a7-sys\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.602383 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/96e55f05-de52-4914-9fc8-7634bff464a7-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.602493 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.602549 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-sys\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.602580 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/96e55f05-de52-4914-9fc8-7634bff464a7-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.602603 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/96e55f05-de52-4914-9fc8-7634bff464a7-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.602648 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96e55f05-de52-4914-9fc8-7634bff464a7-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.602692 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-lib-modules\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.602719 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-run\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.602737 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pg4hj\" (UniqueName: \"kubernetes.io/projected/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-kube-api-access-pg4hj\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.602761 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96e55f05-de52-4914-9fc8-7634bff464a7-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.602815 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/96e55f05-de52-4914-9fc8-7634bff464a7-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.602868 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-ceph\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.603418 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.603471 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/96e55f05-de52-4914-9fc8-7634bff464a7-dev\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.604243 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/96e55f05-de52-4914-9fc8-7634bff464a7-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.605156 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/96e55f05-de52-4914-9fc8-7634bff464a7-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.605364 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-etc-nvme\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.605394 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-lib-modules\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.605414 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/96e55f05-de52-4914-9fc8-7634bff464a7-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.605389 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.605443 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-run\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.605482 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-sys\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.605533 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.605608 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.605610 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/96e55f05-de52-4914-9fc8-7634bff464a7-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.605638 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.605681 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/96e55f05-de52-4914-9fc8-7634bff464a7-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.605720 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-dev\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.605754 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/96e55f05-de52-4914-9fc8-7634bff464a7-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.605834 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/96e55f05-de52-4914-9fc8-7634bff464a7-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.605898 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/96e55f05-de52-4914-9fc8-7634bff464a7-sys\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.610812 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-config-data\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.617698 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-config-data-custom\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.617999 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-ceph\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.618222 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-scripts\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.618614 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/96e55f05-de52-4914-9fc8-7634bff464a7-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.620193 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwnhq\" (UniqueName: \"kubernetes.io/projected/96e55f05-de52-4914-9fc8-7634bff464a7-kube-api-access-kwnhq\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.620779 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.620836 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/96e55f05-de52-4914-9fc8-7634bff464a7-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.621618 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pg4hj\" (UniqueName: \"kubernetes.io/projected/3ba752bd-4b19-4ac4-9798-60de8b5a6f3e-kube-api-access-pg4hj\") pod \"cinder-backup-0\" (UID: \"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e\") " pod="openstack/cinder-backup-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.622888 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96e55f05-de52-4914-9fc8-7634bff464a7-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.625993 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96e55f05-de52-4914-9fc8-7634bff464a7-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.627549 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96e55f05-de52-4914-9fc8-7634bff464a7-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"96e55f05-de52-4914-9fc8-7634bff464a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.761325 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:45 crc kubenswrapper[4873]: I1201 09:32:45.819097 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.123545 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-create-v7b4d"] Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.125689 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-v7b4d" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.139306 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7bd6859fdf-xwffw"] Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.141504 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7bd6859fdf-xwffw" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.145281 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-v7b4d"] Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.147895 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-jmqb9" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.148125 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.148230 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.148350 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.162609 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7bd6859fdf-xwffw"] Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.235079 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8a98544-560c-487a-9a71-59b53a882c9b-operator-scripts\") pod \"manila-db-create-v7b4d\" (UID: \"a8a98544-560c-487a-9a71-59b53a882c9b\") " pod="openstack/manila-db-create-v7b4d" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.235315 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a963ef99-5387-4db3-bbe3-b3cedcdb899d-horizon-secret-key\") pod \"horizon-7bd6859fdf-xwffw\" (UID: \"a963ef99-5387-4db3-bbe3-b3cedcdb899d\") " pod="openstack/horizon-7bd6859fdf-xwffw" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.235380 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a963ef99-5387-4db3-bbe3-b3cedcdb899d-scripts\") pod \"horizon-7bd6859fdf-xwffw\" (UID: \"a963ef99-5387-4db3-bbe3-b3cedcdb899d\") " pod="openstack/horizon-7bd6859fdf-xwffw" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.235503 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a963ef99-5387-4db3-bbe3-b3cedcdb899d-logs\") pod \"horizon-7bd6859fdf-xwffw\" (UID: \"a963ef99-5387-4db3-bbe3-b3cedcdb899d\") " pod="openstack/horizon-7bd6859fdf-xwffw" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.235590 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzxnb\" (UniqueName: \"kubernetes.io/projected/a963ef99-5387-4db3-bbe3-b3cedcdb899d-kube-api-access-dzxnb\") pod \"horizon-7bd6859fdf-xwffw\" (UID: \"a963ef99-5387-4db3-bbe3-b3cedcdb899d\") " pod="openstack/horizon-7bd6859fdf-xwffw" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.235631 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8r9c5\" (UniqueName: \"kubernetes.io/projected/a8a98544-560c-487a-9a71-59b53a882c9b-kube-api-access-8r9c5\") pod \"manila-db-create-v7b4d\" (UID: \"a8a98544-560c-487a-9a71-59b53a882c9b\") " pod="openstack/manila-db-create-v7b4d" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.235693 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a963ef99-5387-4db3-bbe3-b3cedcdb899d-config-data\") pod \"horizon-7bd6859fdf-xwffw\" (UID: \"a963ef99-5387-4db3-bbe3-b3cedcdb899d\") " pod="openstack/horizon-7bd6859fdf-xwffw" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.259394 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.261751 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.264789 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-jmjhl" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.264987 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.266256 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.268189 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.296347 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-9a52-account-create-update-vtg5d"] Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.297941 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-9a52-account-create-update-vtg5d" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.305277 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.307203 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-db-secret" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.335142 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-9a52-account-create-update-vtg5d"] Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.354953 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8r9c5\" (UniqueName: \"kubernetes.io/projected/a8a98544-560c-487a-9a71-59b53a882c9b-kube-api-access-8r9c5\") pod \"manila-db-create-v7b4d\" (UID: \"a8a98544-560c-487a-9a71-59b53a882c9b\") " pod="openstack/manila-db-create-v7b4d" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.355031 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a963ef99-5387-4db3-bbe3-b3cedcdb899d-config-data\") pod \"horizon-7bd6859fdf-xwffw\" (UID: \"a963ef99-5387-4db3-bbe3-b3cedcdb899d\") " pod="openstack/horizon-7bd6859fdf-xwffw" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.355061 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmzqn\" (UniqueName: \"kubernetes.io/projected/483831f0-8b26-4845-b893-50bb7fb6ddb4-kube-api-access-vmzqn\") pod \"manila-9a52-account-create-update-vtg5d\" (UID: \"483831f0-8b26-4845-b893-50bb7fb6ddb4\") " pod="openstack/manila-9a52-account-create-update-vtg5d" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.355118 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/483831f0-8b26-4845-b893-50bb7fb6ddb4-operator-scripts\") pod \"manila-9a52-account-create-update-vtg5d\" (UID: \"483831f0-8b26-4845-b893-50bb7fb6ddb4\") " pod="openstack/manila-9a52-account-create-update-vtg5d" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.355144 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8a98544-560c-487a-9a71-59b53a882c9b-operator-scripts\") pod \"manila-db-create-v7b4d\" (UID: \"a8a98544-560c-487a-9a71-59b53a882c9b\") " pod="openstack/manila-db-create-v7b4d" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.355197 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a963ef99-5387-4db3-bbe3-b3cedcdb899d-horizon-secret-key\") pod \"horizon-7bd6859fdf-xwffw\" (UID: \"a963ef99-5387-4db3-bbe3-b3cedcdb899d\") " pod="openstack/horizon-7bd6859fdf-xwffw" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.355225 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a963ef99-5387-4db3-bbe3-b3cedcdb899d-scripts\") pod \"horizon-7bd6859fdf-xwffw\" (UID: \"a963ef99-5387-4db3-bbe3-b3cedcdb899d\") " pod="openstack/horizon-7bd6859fdf-xwffw" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.355261 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a963ef99-5387-4db3-bbe3-b3cedcdb899d-logs\") pod \"horizon-7bd6859fdf-xwffw\" (UID: \"a963ef99-5387-4db3-bbe3-b3cedcdb899d\") " pod="openstack/horizon-7bd6859fdf-xwffw" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.355288 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzxnb\" (UniqueName: \"kubernetes.io/projected/a963ef99-5387-4db3-bbe3-b3cedcdb899d-kube-api-access-dzxnb\") pod \"horizon-7bd6859fdf-xwffw\" (UID: \"a963ef99-5387-4db3-bbe3-b3cedcdb899d\") " pod="openstack/horizon-7bd6859fdf-xwffw" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.356669 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8a98544-560c-487a-9a71-59b53a882c9b-operator-scripts\") pod \"manila-db-create-v7b4d\" (UID: \"a8a98544-560c-487a-9a71-59b53a882c9b\") " pod="openstack/manila-db-create-v7b4d" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.360291 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a963ef99-5387-4db3-bbe3-b3cedcdb899d-config-data\") pod \"horizon-7bd6859fdf-xwffw\" (UID: \"a963ef99-5387-4db3-bbe3-b3cedcdb899d\") " pod="openstack/horizon-7bd6859fdf-xwffw" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.362076 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.362534 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a963ef99-5387-4db3-bbe3-b3cedcdb899d-logs\") pod \"horizon-7bd6859fdf-xwffw\" (UID: \"a963ef99-5387-4db3-bbe3-b3cedcdb899d\") " pod="openstack/horizon-7bd6859fdf-xwffw" Dec 01 09:32:46 crc kubenswrapper[4873]: E1201 09:32:46.363225 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[ceph combined-ca-bundle config-data glance httpd-run kube-api-access-zrzt9 logs public-tls-certs scripts], unattached volumes=[], failed to process volumes=[ceph combined-ca-bundle config-data glance httpd-run kube-api-access-zrzt9 logs public-tls-certs scripts]: context canceled" pod="openstack/glance-default-external-api-0" podUID="8c578ea2-6ed7-48e3-a50c-11c8b0d644f5" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.373993 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a963ef99-5387-4db3-bbe3-b3cedcdb899d-horizon-secret-key\") pod \"horizon-7bd6859fdf-xwffw\" (UID: \"a963ef99-5387-4db3-bbe3-b3cedcdb899d\") " pod="openstack/horizon-7bd6859fdf-xwffw" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.381112 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a963ef99-5387-4db3-bbe3-b3cedcdb899d-scripts\") pod \"horizon-7bd6859fdf-xwffw\" (UID: \"a963ef99-5387-4db3-bbe3-b3cedcdb899d\") " pod="openstack/horizon-7bd6859fdf-xwffw" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.381242 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.383445 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.385573 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzxnb\" (UniqueName: \"kubernetes.io/projected/a963ef99-5387-4db3-bbe3-b3cedcdb899d-kube-api-access-dzxnb\") pod \"horizon-7bd6859fdf-xwffw\" (UID: \"a963ef99-5387-4db3-bbe3-b3cedcdb899d\") " pod="openstack/horizon-7bd6859fdf-xwffw" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.386057 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8r9c5\" (UniqueName: \"kubernetes.io/projected/a8a98544-560c-487a-9a71-59b53a882c9b-kube-api-access-8r9c5\") pod \"manila-db-create-v7b4d\" (UID: \"a8a98544-560c-487a-9a71-59b53a882c9b\") " pod="openstack/manila-db-create-v7b4d" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.388128 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.392126 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.434230 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.459526 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrzt9\" (UniqueName: \"kubernetes.io/projected/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-kube-api-access-zrzt9\") pod \"glance-default-external-api-0\" (UID: \"8c578ea2-6ed7-48e3-a50c-11c8b0d644f5\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.459611 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-scripts\") pod \"glance-default-external-api-0\" (UID: \"8c578ea2-6ed7-48e3-a50c-11c8b0d644f5\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.459700 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"8c578ea2-6ed7-48e3-a50c-11c8b0d644f5\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.459728 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8c578ea2-6ed7-48e3-a50c-11c8b0d644f5\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.459772 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmzqn\" (UniqueName: \"kubernetes.io/projected/483831f0-8b26-4845-b893-50bb7fb6ddb4-kube-api-access-vmzqn\") pod \"manila-9a52-account-create-update-vtg5d\" (UID: \"483831f0-8b26-4845-b893-50bb7fb6ddb4\") " pod="openstack/manila-9a52-account-create-update-vtg5d" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.459805 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8c578ea2-6ed7-48e3-a50c-11c8b0d644f5\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.459842 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-logs\") pod \"glance-default-external-api-0\" (UID: \"8c578ea2-6ed7-48e3-a50c-11c8b0d644f5\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.459874 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-config-data\") pod \"glance-default-external-api-0\" (UID: \"8c578ea2-6ed7-48e3-a50c-11c8b0d644f5\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.459921 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/483831f0-8b26-4845-b893-50bb7fb6ddb4-operator-scripts\") pod \"manila-9a52-account-create-update-vtg5d\" (UID: \"483831f0-8b26-4845-b893-50bb7fb6ddb4\") " pod="openstack/manila-9a52-account-create-update-vtg5d" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.459957 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-ceph\") pod \"glance-default-external-api-0\" (UID: \"8c578ea2-6ed7-48e3-a50c-11c8b0d644f5\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.460654 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8c578ea2-6ed7-48e3-a50c-11c8b0d644f5\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.460804 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/483831f0-8b26-4845-b893-50bb7fb6ddb4-operator-scripts\") pod \"manila-9a52-account-create-update-vtg5d\" (UID: \"483831f0-8b26-4845-b893-50bb7fb6ddb4\") " pod="openstack/manila-9a52-account-create-update-vtg5d" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.473250 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-v7b4d" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.498110 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7bd6859fdf-xwffw" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.531066 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmzqn\" (UniqueName: \"kubernetes.io/projected/483831f0-8b26-4845-b893-50bb7fb6ddb4-kube-api-access-vmzqn\") pod \"manila-9a52-account-create-update-vtg5d\" (UID: \"483831f0-8b26-4845-b893-50bb7fb6ddb4\") " pod="openstack/manila-9a52-account-create-update-vtg5d" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.531170 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-66d489fdc7-jqfsv"] Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.533490 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66d489fdc7-jqfsv" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.563055 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8dabd96f-ae8a-46ff-96e0-9b7fcab3567f\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.563200 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-logs\") pod \"glance-default-internal-api-0\" (UID: \"8dabd96f-ae8a-46ff-96e0-9b7fcab3567f\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.563261 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"8c578ea2-6ed7-48e3-a50c-11c8b0d644f5\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.563292 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8c578ea2-6ed7-48e3-a50c-11c8b0d644f5\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.563328 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8dabd96f-ae8a-46ff-96e0-9b7fcab3567f\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.563365 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8c578ea2-6ed7-48e3-a50c-11c8b0d644f5\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.563426 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-logs\") pod \"glance-default-external-api-0\" (UID: \"8c578ea2-6ed7-48e3-a50c-11c8b0d644f5\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.563454 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"8dabd96f-ae8a-46ff-96e0-9b7fcab3567f\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.563507 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-config-data\") pod \"glance-default-external-api-0\" (UID: \"8c578ea2-6ed7-48e3-a50c-11c8b0d644f5\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.563593 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-ceph\") pod \"glance-default-internal-api-0\" (UID: \"8dabd96f-ae8a-46ff-96e0-9b7fcab3567f\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.563619 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8dabd96f-ae8a-46ff-96e0-9b7fcab3567f\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.563642 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"8dabd96f-ae8a-46ff-96e0-9b7fcab3567f\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.563664 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-ceph\") pod \"glance-default-external-api-0\" (UID: \"8c578ea2-6ed7-48e3-a50c-11c8b0d644f5\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.563710 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"8dabd96f-ae8a-46ff-96e0-9b7fcab3567f\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.563742 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8c578ea2-6ed7-48e3-a50c-11c8b0d644f5\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.563798 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrzt9\" (UniqueName: \"kubernetes.io/projected/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-kube-api-access-zrzt9\") pod \"glance-default-external-api-0\" (UID: \"8c578ea2-6ed7-48e3-a50c-11c8b0d644f5\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.563836 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-scripts\") pod \"glance-default-external-api-0\" (UID: \"8c578ea2-6ed7-48e3-a50c-11c8b0d644f5\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.563862 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkwcr\" (UniqueName: \"kubernetes.io/projected/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-kube-api-access-pkwcr\") pod \"glance-default-internal-api-0\" (UID: \"8dabd96f-ae8a-46ff-96e0-9b7fcab3567f\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.568141 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8c578ea2-6ed7-48e3-a50c-11c8b0d644f5\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.568890 4873 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"8c578ea2-6ed7-48e3-a50c-11c8b0d644f5\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-external-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.578031 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8c578ea2-6ed7-48e3-a50c-11c8b0d644f5\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.633410 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-ceph\") pod \"glance-default-external-api-0\" (UID: \"8c578ea2-6ed7-48e3-a50c-11c8b0d644f5\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.675067 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 09:32:46 crc kubenswrapper[4873]: E1201 09:32:46.676254 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[ceph combined-ca-bundle config-data glance httpd-run internal-tls-certs kube-api-access-pkwcr logs scripts], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/glance-default-internal-api-0" podUID="8dabd96f-ae8a-46ff-96e0-9b7fcab3567f" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.678644 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8dabd96f-ae8a-46ff-96e0-9b7fcab3567f\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.678738 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-logs\") pod \"glance-default-internal-api-0\" (UID: \"8dabd96f-ae8a-46ff-96e0-9b7fcab3567f\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.678772 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/76d98e93-51f4-49f0-93cd-6564618e6304-config-data\") pod \"horizon-66d489fdc7-jqfsv\" (UID: \"76d98e93-51f4-49f0-93cd-6564618e6304\") " pod="openstack/horizon-66d489fdc7-jqfsv" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.678815 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/76d98e93-51f4-49f0-93cd-6564618e6304-horizon-secret-key\") pod \"horizon-66d489fdc7-jqfsv\" (UID: \"76d98e93-51f4-49f0-93cd-6564618e6304\") " pod="openstack/horizon-66d489fdc7-jqfsv" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.678847 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8dabd96f-ae8a-46ff-96e0-9b7fcab3567f\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.678903 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"8dabd96f-ae8a-46ff-96e0-9b7fcab3567f\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.678971 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-ceph\") pod \"glance-default-internal-api-0\" (UID: \"8dabd96f-ae8a-46ff-96e0-9b7fcab3567f\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.678990 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8dabd96f-ae8a-46ff-96e0-9b7fcab3567f\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.679378 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-logs\") pod \"glance-default-internal-api-0\" (UID: \"8dabd96f-ae8a-46ff-96e0-9b7fcab3567f\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.680299 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"8dabd96f-ae8a-46ff-96e0-9b7fcab3567f\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.681032 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-9a52-account-create-update-vtg5d" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.682456 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"8dabd96f-ae8a-46ff-96e0-9b7fcab3567f\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.682496 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9tpcf\" (UniqueName: \"kubernetes.io/projected/76d98e93-51f4-49f0-93cd-6564618e6304-kube-api-access-9tpcf\") pod \"horizon-66d489fdc7-jqfsv\" (UID: \"76d98e93-51f4-49f0-93cd-6564618e6304\") " pod="openstack/horizon-66d489fdc7-jqfsv" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.682539 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76d98e93-51f4-49f0-93cd-6564618e6304-logs\") pod \"horizon-66d489fdc7-jqfsv\" (UID: \"76d98e93-51f4-49f0-93cd-6564618e6304\") " pod="openstack/horizon-66d489fdc7-jqfsv" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.682620 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/76d98e93-51f4-49f0-93cd-6564618e6304-scripts\") pod \"horizon-66d489fdc7-jqfsv\" (UID: \"76d98e93-51f4-49f0-93cd-6564618e6304\") " pod="openstack/horizon-66d489fdc7-jqfsv" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.682625 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8dabd96f-ae8a-46ff-96e0-9b7fcab3567f\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.682691 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkwcr\" (UniqueName: \"kubernetes.io/projected/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-kube-api-access-pkwcr\") pod \"glance-default-internal-api-0\" (UID: \"8dabd96f-ae8a-46ff-96e0-9b7fcab3567f\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.682709 4873 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"8dabd96f-ae8a-46ff-96e0-9b7fcab3567f\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-internal-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.684549 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-scripts\") pod \"glance-default-external-api-0\" (UID: \"8c578ea2-6ed7-48e3-a50c-11c8b0d644f5\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.698355 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-66d489fdc7-jqfsv"] Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.700506 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-config-data\") pod \"glance-default-external-api-0\" (UID: \"8c578ea2-6ed7-48e3-a50c-11c8b0d644f5\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.700713 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"8dabd96f-ae8a-46ff-96e0-9b7fcab3567f\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.725038 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkwcr\" (UniqueName: \"kubernetes.io/projected/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-kube-api-access-pkwcr\") pod \"glance-default-internal-api-0\" (UID: \"8dabd96f-ae8a-46ff-96e0-9b7fcab3567f\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.735541 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-logs\") pod \"glance-default-external-api-0\" (UID: \"8c578ea2-6ed7-48e3-a50c-11c8b0d644f5\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.736570 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrzt9\" (UniqueName: \"kubernetes.io/projected/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-kube-api-access-zrzt9\") pod \"glance-default-external-api-0\" (UID: \"8c578ea2-6ed7-48e3-a50c-11c8b0d644f5\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.748005 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8dabd96f-ae8a-46ff-96e0-9b7fcab3567f\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.748352 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-ceph\") pod \"glance-default-internal-api-0\" (UID: \"8dabd96f-ae8a-46ff-96e0-9b7fcab3567f\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.748574 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8dabd96f-ae8a-46ff-96e0-9b7fcab3567f\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.750956 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8c578ea2-6ed7-48e3-a50c-11c8b0d644f5\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.787656 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/76d98e93-51f4-49f0-93cd-6564618e6304-scripts\") pod \"horizon-66d489fdc7-jqfsv\" (UID: \"76d98e93-51f4-49f0-93cd-6564618e6304\") " pod="openstack/horizon-66d489fdc7-jqfsv" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.787822 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/76d98e93-51f4-49f0-93cd-6564618e6304-config-data\") pod \"horizon-66d489fdc7-jqfsv\" (UID: \"76d98e93-51f4-49f0-93cd-6564618e6304\") " pod="openstack/horizon-66d489fdc7-jqfsv" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.787872 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/76d98e93-51f4-49f0-93cd-6564618e6304-horizon-secret-key\") pod \"horizon-66d489fdc7-jqfsv\" (UID: \"76d98e93-51f4-49f0-93cd-6564618e6304\") " pod="openstack/horizon-66d489fdc7-jqfsv" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.788035 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9tpcf\" (UniqueName: \"kubernetes.io/projected/76d98e93-51f4-49f0-93cd-6564618e6304-kube-api-access-9tpcf\") pod \"horizon-66d489fdc7-jqfsv\" (UID: \"76d98e93-51f4-49f0-93cd-6564618e6304\") " pod="openstack/horizon-66d489fdc7-jqfsv" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.788077 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76d98e93-51f4-49f0-93cd-6564618e6304-logs\") pod \"horizon-66d489fdc7-jqfsv\" (UID: \"76d98e93-51f4-49f0-93cd-6564618e6304\") " pod="openstack/horizon-66d489fdc7-jqfsv" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.788807 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76d98e93-51f4-49f0-93cd-6564618e6304-logs\") pod \"horizon-66d489fdc7-jqfsv\" (UID: \"76d98e93-51f4-49f0-93cd-6564618e6304\") " pod="openstack/horizon-66d489fdc7-jqfsv" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.790686 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/76d98e93-51f4-49f0-93cd-6564618e6304-scripts\") pod \"horizon-66d489fdc7-jqfsv\" (UID: \"76d98e93-51f4-49f0-93cd-6564618e6304\") " pod="openstack/horizon-66d489fdc7-jqfsv" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.791928 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/76d98e93-51f4-49f0-93cd-6564618e6304-config-data\") pod \"horizon-66d489fdc7-jqfsv\" (UID: \"76d98e93-51f4-49f0-93cd-6564618e6304\") " pod="openstack/horizon-66d489fdc7-jqfsv" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.793812 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"8dabd96f-ae8a-46ff-96e0-9b7fcab3567f\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.808704 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/76d98e93-51f4-49f0-93cd-6564618e6304-horizon-secret-key\") pod \"horizon-66d489fdc7-jqfsv\" (UID: \"76d98e93-51f4-49f0-93cd-6564618e6304\") " pod="openstack/horizon-66d489fdc7-jqfsv" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.824056 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.929767 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"8c578ea2-6ed7-48e3-a50c-11c8b0d644f5\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.933909 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9tpcf\" (UniqueName: \"kubernetes.io/projected/76d98e93-51f4-49f0-93cd-6564618e6304-kube-api-access-9tpcf\") pod \"horizon-66d489fdc7-jqfsv\" (UID: \"76d98e93-51f4-49f0-93cd-6564618e6304\") " pod="openstack/horizon-66d489fdc7-jqfsv" Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.980198 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Dec 01 09:32:46 crc kubenswrapper[4873]: I1201 09:32:46.986810 4873 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.018785 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"8dabd96f-ae8a-46ff-96e0-9b7fcab3567f\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.051679 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.051958 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"96e55f05-de52-4914-9fc8-7634bff464a7","Type":"ContainerStarted","Data":"4f1da187010fdba35bda61361c25439021db8229034d90e06830e5c9da8d79b4"} Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.053192 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.102332 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66d489fdc7-jqfsv" Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.101273 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.167197 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.245221 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-logs\") pod \"8dabd96f-ae8a-46ff-96e0-9b7fcab3567f\" (UID: \"8dabd96f-ae8a-46ff-96e0-9b7fcab3567f\") " Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.245309 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-httpd-run\") pod \"8dabd96f-ae8a-46ff-96e0-9b7fcab3567f\" (UID: \"8dabd96f-ae8a-46ff-96e0-9b7fcab3567f\") " Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.245392 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-internal-tls-certs\") pod \"8dabd96f-ae8a-46ff-96e0-9b7fcab3567f\" (UID: \"8dabd96f-ae8a-46ff-96e0-9b7fcab3567f\") " Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.245435 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-combined-ca-bundle\") pod \"8dabd96f-ae8a-46ff-96e0-9b7fcab3567f\" (UID: \"8dabd96f-ae8a-46ff-96e0-9b7fcab3567f\") " Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.245452 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-scripts\") pod \"8dabd96f-ae8a-46ff-96e0-9b7fcab3567f\" (UID: \"8dabd96f-ae8a-46ff-96e0-9b7fcab3567f\") " Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.245548 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"8dabd96f-ae8a-46ff-96e0-9b7fcab3567f\" (UID: \"8dabd96f-ae8a-46ff-96e0-9b7fcab3567f\") " Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.245589 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-ceph\") pod \"8dabd96f-ae8a-46ff-96e0-9b7fcab3567f\" (UID: \"8dabd96f-ae8a-46ff-96e0-9b7fcab3567f\") " Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.245621 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-config-data\") pod \"8dabd96f-ae8a-46ff-96e0-9b7fcab3567f\" (UID: \"8dabd96f-ae8a-46ff-96e0-9b7fcab3567f\") " Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.245652 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pkwcr\" (UniqueName: \"kubernetes.io/projected/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-kube-api-access-pkwcr\") pod \"8dabd96f-ae8a-46ff-96e0-9b7fcab3567f\" (UID: \"8dabd96f-ae8a-46ff-96e0-9b7fcab3567f\") " Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.250449 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-logs" (OuterVolumeSpecName: "logs") pod "8dabd96f-ae8a-46ff-96e0-9b7fcab3567f" (UID: "8dabd96f-ae8a-46ff-96e0-9b7fcab3567f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.250484 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "8dabd96f-ae8a-46ff-96e0-9b7fcab3567f" (UID: "8dabd96f-ae8a-46ff-96e0-9b7fcab3567f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.253438 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "8dabd96f-ae8a-46ff-96e0-9b7fcab3567f" (UID: "8dabd96f-ae8a-46ff-96e0-9b7fcab3567f"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.254352 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-kube-api-access-pkwcr" (OuterVolumeSpecName: "kube-api-access-pkwcr") pod "8dabd96f-ae8a-46ff-96e0-9b7fcab3567f" (UID: "8dabd96f-ae8a-46ff-96e0-9b7fcab3567f"). InnerVolumeSpecName "kube-api-access-pkwcr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.255428 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-ceph" (OuterVolumeSpecName: "ceph") pod "8dabd96f-ae8a-46ff-96e0-9b7fcab3567f" (UID: "8dabd96f-ae8a-46ff-96e0-9b7fcab3567f"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.257287 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8dabd96f-ae8a-46ff-96e0-9b7fcab3567f" (UID: "8dabd96f-ae8a-46ff-96e0-9b7fcab3567f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.257338 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "8dabd96f-ae8a-46ff-96e0-9b7fcab3567f" (UID: "8dabd96f-ae8a-46ff-96e0-9b7fcab3567f"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.258506 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-config-data" (OuterVolumeSpecName: "config-data") pod "8dabd96f-ae8a-46ff-96e0-9b7fcab3567f" (UID: "8dabd96f-ae8a-46ff-96e0-9b7fcab3567f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.268185 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-scripts" (OuterVolumeSpecName: "scripts") pod "8dabd96f-ae8a-46ff-96e0-9b7fcab3567f" (UID: "8dabd96f-ae8a-46ff-96e0-9b7fcab3567f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.347837 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zrzt9\" (UniqueName: \"kubernetes.io/projected/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-kube-api-access-zrzt9\") pod \"8c578ea2-6ed7-48e3-a50c-11c8b0d644f5\" (UID: \"8c578ea2-6ed7-48e3-a50c-11c8b0d644f5\") " Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.347943 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-logs\") pod \"8c578ea2-6ed7-48e3-a50c-11c8b0d644f5\" (UID: \"8c578ea2-6ed7-48e3-a50c-11c8b0d644f5\") " Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.348023 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-config-data\") pod \"8c578ea2-6ed7-48e3-a50c-11c8b0d644f5\" (UID: \"8c578ea2-6ed7-48e3-a50c-11c8b0d644f5\") " Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.348254 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-combined-ca-bundle\") pod \"8c578ea2-6ed7-48e3-a50c-11c8b0d644f5\" (UID: \"8c578ea2-6ed7-48e3-a50c-11c8b0d644f5\") " Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.348358 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-public-tls-certs\") pod \"8c578ea2-6ed7-48e3-a50c-11c8b0d644f5\" (UID: \"8c578ea2-6ed7-48e3-a50c-11c8b0d644f5\") " Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.348405 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-ceph\") pod \"8c578ea2-6ed7-48e3-a50c-11c8b0d644f5\" (UID: \"8c578ea2-6ed7-48e3-a50c-11c8b0d644f5\") " Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.348465 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-httpd-run\") pod \"8c578ea2-6ed7-48e3-a50c-11c8b0d644f5\" (UID: \"8c578ea2-6ed7-48e3-a50c-11c8b0d644f5\") " Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.348493 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-scripts\") pod \"8c578ea2-6ed7-48e3-a50c-11c8b0d644f5\" (UID: \"8c578ea2-6ed7-48e3-a50c-11c8b0d644f5\") " Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.348593 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"8c578ea2-6ed7-48e3-a50c-11c8b0d644f5\" (UID: \"8c578ea2-6ed7-48e3-a50c-11c8b0d644f5\") " Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.349317 4873 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-logs\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.349342 4873 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.349354 4873 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.349369 4873 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.349381 4873 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.349409 4873 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.349420 4873 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-ceph\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.349434 4873 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.349446 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pkwcr\" (UniqueName: \"kubernetes.io/projected/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f-kube-api-access-pkwcr\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.357187 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "8c578ea2-6ed7-48e3-a50c-11c8b0d644f5" (UID: "8c578ea2-6ed7-48e3-a50c-11c8b0d644f5"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.358639 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-logs" (OuterVolumeSpecName: "logs") pod "8c578ea2-6ed7-48e3-a50c-11c8b0d644f5" (UID: "8c578ea2-6ed7-48e3-a50c-11c8b0d644f5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.359541 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "8c578ea2-6ed7-48e3-a50c-11c8b0d644f5" (UID: "8c578ea2-6ed7-48e3-a50c-11c8b0d644f5"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.361029 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-ceph" (OuterVolumeSpecName: "ceph") pod "8c578ea2-6ed7-48e3-a50c-11c8b0d644f5" (UID: "8c578ea2-6ed7-48e3-a50c-11c8b0d644f5"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.362253 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-scripts" (OuterVolumeSpecName: "scripts") pod "8c578ea2-6ed7-48e3-a50c-11c8b0d644f5" (UID: "8c578ea2-6ed7-48e3-a50c-11c8b0d644f5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.364408 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "8c578ea2-6ed7-48e3-a50c-11c8b0d644f5" (UID: "8c578ea2-6ed7-48e3-a50c-11c8b0d644f5"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.364539 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8c578ea2-6ed7-48e3-a50c-11c8b0d644f5" (UID: "8c578ea2-6ed7-48e3-a50c-11c8b0d644f5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.364667 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-config-data" (OuterVolumeSpecName: "config-data") pod "8c578ea2-6ed7-48e3-a50c-11c8b0d644f5" (UID: "8c578ea2-6ed7-48e3-a50c-11c8b0d644f5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.367631 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-kube-api-access-zrzt9" (OuterVolumeSpecName: "kube-api-access-zrzt9") pod "8c578ea2-6ed7-48e3-a50c-11c8b0d644f5" (UID: "8c578ea2-6ed7-48e3-a50c-11c8b0d644f5"). InnerVolumeSpecName "kube-api-access-zrzt9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.415158 4873 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.452352 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zrzt9\" (UniqueName: \"kubernetes.io/projected/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-kube-api-access-zrzt9\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.452803 4873 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-logs\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.452821 4873 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.452834 4873 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.452847 4873 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.452859 4873 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.452896 4873 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-ceph\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.452909 4873 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.452923 4873 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.452965 4873 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.492107 4873 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.545317 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-9a52-account-create-update-vtg5d"] Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.555955 4873 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.563058 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-v7b4d"] Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.769688 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-66d489fdc7-jqfsv"] Dec 01 09:32:47 crc kubenswrapper[4873]: I1201 09:32:47.783483 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7bd6859fdf-xwffw"] Dec 01 09:32:47 crc kubenswrapper[4873]: W1201 09:32:47.791753 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda963ef99_5387_4db3_bbe3_b3cedcdb899d.slice/crio-6e79504426d789af3cbc7fb0dd6a12879c634e14e55bbe6673b0511f670a845f WatchSource:0}: Error finding container 6e79504426d789af3cbc7fb0dd6a12879c634e14e55bbe6673b0511f670a845f: Status 404 returned error can't find the container with id 6e79504426d789af3cbc7fb0dd6a12879c634e14e55bbe6673b0511f670a845f Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.071287 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66d489fdc7-jqfsv" event={"ID":"76d98e93-51f4-49f0-93cd-6564618e6304","Type":"ContainerStarted","Data":"450b355ea6db79f14f51d471535dffa9a11e3538ff03d5dbd44c56063bb965b0"} Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.077041 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-v7b4d" event={"ID":"a8a98544-560c-487a-9a71-59b53a882c9b","Type":"ContainerStarted","Data":"d351090aab0a1519656e52e7efd5da92255a58fcd1e22b6940ce50fac5ff0e41"} Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.077089 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-v7b4d" event={"ID":"a8a98544-560c-487a-9a71-59b53a882c9b","Type":"ContainerStarted","Data":"492c39da1084067e6b717eb2043de864da8215bd57280dd57d132a0499eb4b88"} Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.084084 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e","Type":"ContainerStarted","Data":"e48531b0dd9a54b5c9bf2713fb25ab5d7e8ecb04a33d4c7d1934ba1530a4c037"} Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.089476 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bd6859fdf-xwffw" event={"ID":"a963ef99-5387-4db3-bbe3-b3cedcdb899d","Type":"ContainerStarted","Data":"6e79504426d789af3cbc7fb0dd6a12879c634e14e55bbe6673b0511f670a845f"} Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.092062 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.092341 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-9a52-account-create-update-vtg5d" event={"ID":"483831f0-8b26-4845-b893-50bb7fb6ddb4","Type":"ContainerStarted","Data":"0e27232c827dbdbca9e3aed00ce658d6ef88d4774e4add85cdbbbdef58bdbc52"} Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.092395 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-9a52-account-create-update-vtg5d" event={"ID":"483831f0-8b26-4845-b893-50bb7fb6ddb4","Type":"ContainerStarted","Data":"fa31785d87239fe681ce10febf6e012eccd78ea7adf94b2285a244e3e27c8a6c"} Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.092459 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.102142 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-create-v7b4d" podStartSLOduration=2.102120037 podStartE2EDuration="2.102120037s" podCreationTimestamp="2025-12-01 09:32:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:32:48.09743267 +0000 UTC m=+3143.999541209" watchObservedRunningTime="2025-12-01 09:32:48.102120037 +0000 UTC m=+3144.004228566" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.134340 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-9a52-account-create-update-vtg5d" podStartSLOduration=2.134318793 podStartE2EDuration="2.134318793s" podCreationTimestamp="2025-12-01 09:32:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:32:48.128524318 +0000 UTC m=+3144.030632857" watchObservedRunningTime="2025-12-01 09:32:48.134318793 +0000 UTC m=+3144.036427332" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.248491 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.271800 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.302102 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.323042 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.331596 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.335519 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-jmjhl" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.341078 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.341397 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.341528 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.352111 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.357507 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.370130 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.392277 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.392451 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.399153 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.400581 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.476417 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c578ea2-6ed7-48e3-a50c-11c8b0d644f5" path="/var/lib/kubelet/pods/8c578ea2-6ed7-48e3-a50c-11c8b0d644f5/volumes" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.477207 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8dabd96f-ae8a-46ff-96e0-9b7fcab3567f" path="/var/lib/kubelet/pods/8dabd96f-ae8a-46ff-96e0-9b7fcab3567f/volumes" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.478715 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5dccf2f5-f149-4b29-99a5-560ff38df6f1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5dccf2f5-f149-4b29-99a5-560ff38df6f1\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.478776 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/354e9917-810a-44ba-bc85-b12e15152129-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"354e9917-810a-44ba-bc85-b12e15152129\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.478823 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/354e9917-810a-44ba-bc85-b12e15152129-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"354e9917-810a-44ba-bc85-b12e15152129\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.478850 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/354e9917-810a-44ba-bc85-b12e15152129-config-data\") pod \"glance-default-external-api-0\" (UID: \"354e9917-810a-44ba-bc85-b12e15152129\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.478892 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkmtw\" (UniqueName: \"kubernetes.io/projected/354e9917-810a-44ba-bc85-b12e15152129-kube-api-access-rkmtw\") pod \"glance-default-external-api-0\" (UID: \"354e9917-810a-44ba-bc85-b12e15152129\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.478943 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/354e9917-810a-44ba-bc85-b12e15152129-logs\") pod \"glance-default-external-api-0\" (UID: \"354e9917-810a-44ba-bc85-b12e15152129\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.478963 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5dccf2f5-f149-4b29-99a5-560ff38df6f1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5dccf2f5-f149-4b29-99a5-560ff38df6f1\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.478980 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htdzz\" (UniqueName: \"kubernetes.io/projected/5dccf2f5-f149-4b29-99a5-560ff38df6f1-kube-api-access-htdzz\") pod \"glance-default-internal-api-0\" (UID: \"5dccf2f5-f149-4b29-99a5-560ff38df6f1\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.479041 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5dccf2f5-f149-4b29-99a5-560ff38df6f1-ceph\") pod \"glance-default-internal-api-0\" (UID: \"5dccf2f5-f149-4b29-99a5-560ff38df6f1\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.479061 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5dccf2f5-f149-4b29-99a5-560ff38df6f1-logs\") pod \"glance-default-internal-api-0\" (UID: \"5dccf2f5-f149-4b29-99a5-560ff38df6f1\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.479080 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/354e9917-810a-44ba-bc85-b12e15152129-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"354e9917-810a-44ba-bc85-b12e15152129\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.479125 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/354e9917-810a-44ba-bc85-b12e15152129-ceph\") pod \"glance-default-external-api-0\" (UID: \"354e9917-810a-44ba-bc85-b12e15152129\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.479147 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"354e9917-810a-44ba-bc85-b12e15152129\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.480260 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5dccf2f5-f149-4b29-99a5-560ff38df6f1-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5dccf2f5-f149-4b29-99a5-560ff38df6f1\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.480377 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"5dccf2f5-f149-4b29-99a5-560ff38df6f1\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.480496 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/354e9917-810a-44ba-bc85-b12e15152129-scripts\") pod \"glance-default-external-api-0\" (UID: \"354e9917-810a-44ba-bc85-b12e15152129\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.480515 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5dccf2f5-f149-4b29-99a5-560ff38df6f1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5dccf2f5-f149-4b29-99a5-560ff38df6f1\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.480587 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5dccf2f5-f149-4b29-99a5-560ff38df6f1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5dccf2f5-f149-4b29-99a5-560ff38df6f1\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.582288 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5dccf2f5-f149-4b29-99a5-560ff38df6f1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5dccf2f5-f149-4b29-99a5-560ff38df6f1\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.583736 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/354e9917-810a-44ba-bc85-b12e15152129-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"354e9917-810a-44ba-bc85-b12e15152129\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.583909 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/354e9917-810a-44ba-bc85-b12e15152129-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"354e9917-810a-44ba-bc85-b12e15152129\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.584173 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/354e9917-810a-44ba-bc85-b12e15152129-config-data\") pod \"glance-default-external-api-0\" (UID: \"354e9917-810a-44ba-bc85-b12e15152129\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.584347 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkmtw\" (UniqueName: \"kubernetes.io/projected/354e9917-810a-44ba-bc85-b12e15152129-kube-api-access-rkmtw\") pod \"glance-default-external-api-0\" (UID: \"354e9917-810a-44ba-bc85-b12e15152129\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.584434 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/354e9917-810a-44ba-bc85-b12e15152129-logs\") pod \"glance-default-external-api-0\" (UID: \"354e9917-810a-44ba-bc85-b12e15152129\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.584508 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5dccf2f5-f149-4b29-99a5-560ff38df6f1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5dccf2f5-f149-4b29-99a5-560ff38df6f1\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.584600 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htdzz\" (UniqueName: \"kubernetes.io/projected/5dccf2f5-f149-4b29-99a5-560ff38df6f1-kube-api-access-htdzz\") pod \"glance-default-internal-api-0\" (UID: \"5dccf2f5-f149-4b29-99a5-560ff38df6f1\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.584729 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5dccf2f5-f149-4b29-99a5-560ff38df6f1-ceph\") pod \"glance-default-internal-api-0\" (UID: \"5dccf2f5-f149-4b29-99a5-560ff38df6f1\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.584855 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5dccf2f5-f149-4b29-99a5-560ff38df6f1-logs\") pod \"glance-default-internal-api-0\" (UID: \"5dccf2f5-f149-4b29-99a5-560ff38df6f1\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.584953 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/354e9917-810a-44ba-bc85-b12e15152129-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"354e9917-810a-44ba-bc85-b12e15152129\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.585416 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/354e9917-810a-44ba-bc85-b12e15152129-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"354e9917-810a-44ba-bc85-b12e15152129\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.585725 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/354e9917-810a-44ba-bc85-b12e15152129-logs\") pod \"glance-default-external-api-0\" (UID: \"354e9917-810a-44ba-bc85-b12e15152129\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.585933 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/354e9917-810a-44ba-bc85-b12e15152129-ceph\") pod \"glance-default-external-api-0\" (UID: \"354e9917-810a-44ba-bc85-b12e15152129\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.585984 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"354e9917-810a-44ba-bc85-b12e15152129\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.586120 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5dccf2f5-f149-4b29-99a5-560ff38df6f1-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5dccf2f5-f149-4b29-99a5-560ff38df6f1\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.586204 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"5dccf2f5-f149-4b29-99a5-560ff38df6f1\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.586241 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/354e9917-810a-44ba-bc85-b12e15152129-scripts\") pod \"glance-default-external-api-0\" (UID: \"354e9917-810a-44ba-bc85-b12e15152129\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.586264 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5dccf2f5-f149-4b29-99a5-560ff38df6f1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5dccf2f5-f149-4b29-99a5-560ff38df6f1\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.586355 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5dccf2f5-f149-4b29-99a5-560ff38df6f1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5dccf2f5-f149-4b29-99a5-560ff38df6f1\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.587158 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5dccf2f5-f149-4b29-99a5-560ff38df6f1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5dccf2f5-f149-4b29-99a5-560ff38df6f1\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.587447 4873 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"354e9917-810a-44ba-bc85-b12e15152129\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-external-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.587507 4873 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"5dccf2f5-f149-4b29-99a5-560ff38df6f1\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-internal-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.588278 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5dccf2f5-f149-4b29-99a5-560ff38df6f1-logs\") pod \"glance-default-internal-api-0\" (UID: \"5dccf2f5-f149-4b29-99a5-560ff38df6f1\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.590068 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/354e9917-810a-44ba-bc85-b12e15152129-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"354e9917-810a-44ba-bc85-b12e15152129\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.590940 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5dccf2f5-f149-4b29-99a5-560ff38df6f1-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5dccf2f5-f149-4b29-99a5-560ff38df6f1\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.590944 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/354e9917-810a-44ba-bc85-b12e15152129-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"354e9917-810a-44ba-bc85-b12e15152129\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.591364 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5dccf2f5-f149-4b29-99a5-560ff38df6f1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5dccf2f5-f149-4b29-99a5-560ff38df6f1\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.595962 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5dccf2f5-f149-4b29-99a5-560ff38df6f1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5dccf2f5-f149-4b29-99a5-560ff38df6f1\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.597004 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/354e9917-810a-44ba-bc85-b12e15152129-config-data\") pod \"glance-default-external-api-0\" (UID: \"354e9917-810a-44ba-bc85-b12e15152129\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.597613 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/354e9917-810a-44ba-bc85-b12e15152129-ceph\") pod \"glance-default-external-api-0\" (UID: \"354e9917-810a-44ba-bc85-b12e15152129\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.606982 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5dccf2f5-f149-4b29-99a5-560ff38df6f1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5dccf2f5-f149-4b29-99a5-560ff38df6f1\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.608057 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5dccf2f5-f149-4b29-99a5-560ff38df6f1-ceph\") pod \"glance-default-internal-api-0\" (UID: \"5dccf2f5-f149-4b29-99a5-560ff38df6f1\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.609075 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkmtw\" (UniqueName: \"kubernetes.io/projected/354e9917-810a-44ba-bc85-b12e15152129-kube-api-access-rkmtw\") pod \"glance-default-external-api-0\" (UID: \"354e9917-810a-44ba-bc85-b12e15152129\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.610286 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/354e9917-810a-44ba-bc85-b12e15152129-scripts\") pod \"glance-default-external-api-0\" (UID: \"354e9917-810a-44ba-bc85-b12e15152129\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.627841 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htdzz\" (UniqueName: \"kubernetes.io/projected/5dccf2f5-f149-4b29-99a5-560ff38df6f1-kube-api-access-htdzz\") pod \"glance-default-internal-api-0\" (UID: \"5dccf2f5-f149-4b29-99a5-560ff38df6f1\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.655946 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"354e9917-810a-44ba-bc85-b12e15152129\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.712452 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"5dccf2f5-f149-4b29-99a5-560ff38df6f1\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.807616 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 09:32:48 crc kubenswrapper[4873]: I1201 09:32:48.823632 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 09:32:49 crc kubenswrapper[4873]: I1201 09:32:49.120553 4873 generic.go:334] "Generic (PLEG): container finished" podID="a8a98544-560c-487a-9a71-59b53a882c9b" containerID="d351090aab0a1519656e52e7efd5da92255a58fcd1e22b6940ce50fac5ff0e41" exitCode=0 Dec 01 09:32:49 crc kubenswrapper[4873]: I1201 09:32:49.120902 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-v7b4d" event={"ID":"a8a98544-560c-487a-9a71-59b53a882c9b","Type":"ContainerDied","Data":"d351090aab0a1519656e52e7efd5da92255a58fcd1e22b6940ce50fac5ff0e41"} Dec 01 09:32:49 crc kubenswrapper[4873]: I1201 09:32:49.124206 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e","Type":"ContainerStarted","Data":"a8f562d599e33e3ada4c1db6ae7a296ab28c86861e2448a9401bcf6b3293b95d"} Dec 01 09:32:49 crc kubenswrapper[4873]: I1201 09:32:49.126589 4873 generic.go:334] "Generic (PLEG): container finished" podID="483831f0-8b26-4845-b893-50bb7fb6ddb4" containerID="0e27232c827dbdbca9e3aed00ce658d6ef88d4774e4add85cdbbbdef58bdbc52" exitCode=0 Dec 01 09:32:49 crc kubenswrapper[4873]: I1201 09:32:49.126640 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-9a52-account-create-update-vtg5d" event={"ID":"483831f0-8b26-4845-b893-50bb7fb6ddb4","Type":"ContainerDied","Data":"0e27232c827dbdbca9e3aed00ce658d6ef88d4774e4add85cdbbbdef58bdbc52"} Dec 01 09:32:49 crc kubenswrapper[4873]: I1201 09:32:49.130599 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"96e55f05-de52-4914-9fc8-7634bff464a7","Type":"ContainerStarted","Data":"22a8e44ec20f82e1e971978e6b9b251e3024fb3571afd96dc524765adadb6b7e"} Dec 01 09:32:49 crc kubenswrapper[4873]: I1201 09:32:49.547593 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 09:32:49 crc kubenswrapper[4873]: W1201 09:32:49.583760 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod354e9917_810a_44ba_bc85_b12e15152129.slice/crio-3dc60723b6966df88ba5c9c02f4bf0307620723e1f9b69392e836c0fe6c0ad8c WatchSource:0}: Error finding container 3dc60723b6966df88ba5c9c02f4bf0307620723e1f9b69392e836c0fe6c0ad8c: Status 404 returned error can't find the container with id 3dc60723b6966df88ba5c9c02f4bf0307620723e1f9b69392e836c0fe6c0ad8c Dec 01 09:32:49 crc kubenswrapper[4873]: I1201 09:32:49.626869 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 09:32:49 crc kubenswrapper[4873]: I1201 09:32:49.890958 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-66d489fdc7-jqfsv"] Dec 01 09:32:49 crc kubenswrapper[4873]: I1201 09:32:49.938032 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 09:32:49 crc kubenswrapper[4873]: I1201 09:32:49.979099 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7c885c5cb8-46fsf"] Dec 01 09:32:49 crc kubenswrapper[4873]: I1201 09:32:49.981320 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c885c5cb8-46fsf" Dec 01 09:32:49 crc kubenswrapper[4873]: I1201 09:32:49.986913 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.008992 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7c885c5cb8-46fsf"] Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.057150 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/24452fe2-b990-4537-b619-9db436ed5acb-horizon-secret-key\") pod \"horizon-7c885c5cb8-46fsf\" (UID: \"24452fe2-b990-4537-b619-9db436ed5acb\") " pod="openstack/horizon-7c885c5cb8-46fsf" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.057255 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/24452fe2-b990-4537-b619-9db436ed5acb-logs\") pod \"horizon-7c885c5cb8-46fsf\" (UID: \"24452fe2-b990-4537-b619-9db436ed5acb\") " pod="openstack/horizon-7c885c5cb8-46fsf" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.057337 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/24452fe2-b990-4537-b619-9db436ed5acb-config-data\") pod \"horizon-7c885c5cb8-46fsf\" (UID: \"24452fe2-b990-4537-b619-9db436ed5acb\") " pod="openstack/horizon-7c885c5cb8-46fsf" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.057373 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/24452fe2-b990-4537-b619-9db436ed5acb-horizon-tls-certs\") pod \"horizon-7c885c5cb8-46fsf\" (UID: \"24452fe2-b990-4537-b619-9db436ed5acb\") " pod="openstack/horizon-7c885c5cb8-46fsf" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.057432 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/24452fe2-b990-4537-b619-9db436ed5acb-scripts\") pod \"horizon-7c885c5cb8-46fsf\" (UID: \"24452fe2-b990-4537-b619-9db436ed5acb\") " pod="openstack/horizon-7c885c5cb8-46fsf" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.057546 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24452fe2-b990-4537-b619-9db436ed5acb-combined-ca-bundle\") pod \"horizon-7c885c5cb8-46fsf\" (UID: \"24452fe2-b990-4537-b619-9db436ed5acb\") " pod="openstack/horizon-7c885c5cb8-46fsf" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.057579 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldm2p\" (UniqueName: \"kubernetes.io/projected/24452fe2-b990-4537-b619-9db436ed5acb-kube-api-access-ldm2p\") pod \"horizon-7c885c5cb8-46fsf\" (UID: \"24452fe2-b990-4537-b619-9db436ed5acb\") " pod="openstack/horizon-7c885c5cb8-46fsf" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.074580 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.085109 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7bd6859fdf-xwffw"] Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.127054 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6968fbf848-w68ch"] Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.129538 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6968fbf848-w68ch" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.161637 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/24452fe2-b990-4537-b619-9db436ed5acb-logs\") pod \"horizon-7c885c5cb8-46fsf\" (UID: \"24452fe2-b990-4537-b619-9db436ed5acb\") " pod="openstack/horizon-7c885c5cb8-46fsf" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.161770 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/24452fe2-b990-4537-b619-9db436ed5acb-config-data\") pod \"horizon-7c885c5cb8-46fsf\" (UID: \"24452fe2-b990-4537-b619-9db436ed5acb\") " pod="openstack/horizon-7c885c5cb8-46fsf" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.161807 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/24452fe2-b990-4537-b619-9db436ed5acb-horizon-tls-certs\") pod \"horizon-7c885c5cb8-46fsf\" (UID: \"24452fe2-b990-4537-b619-9db436ed5acb\") " pod="openstack/horizon-7c885c5cb8-46fsf" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.161866 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/24452fe2-b990-4537-b619-9db436ed5acb-scripts\") pod \"horizon-7c885c5cb8-46fsf\" (UID: \"24452fe2-b990-4537-b619-9db436ed5acb\") " pod="openstack/horizon-7c885c5cb8-46fsf" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.162176 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24452fe2-b990-4537-b619-9db436ed5acb-combined-ca-bundle\") pod \"horizon-7c885c5cb8-46fsf\" (UID: \"24452fe2-b990-4537-b619-9db436ed5acb\") " pod="openstack/horizon-7c885c5cb8-46fsf" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.162230 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldm2p\" (UniqueName: \"kubernetes.io/projected/24452fe2-b990-4537-b619-9db436ed5acb-kube-api-access-ldm2p\") pod \"horizon-7c885c5cb8-46fsf\" (UID: \"24452fe2-b990-4537-b619-9db436ed5acb\") " pod="openstack/horizon-7c885c5cb8-46fsf" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.162265 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/24452fe2-b990-4537-b619-9db436ed5acb-horizon-secret-key\") pod \"horizon-7c885c5cb8-46fsf\" (UID: \"24452fe2-b990-4537-b619-9db436ed5acb\") " pod="openstack/horizon-7c885c5cb8-46fsf" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.170273 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/24452fe2-b990-4537-b619-9db436ed5acb-horizon-secret-key\") pod \"horizon-7c885c5cb8-46fsf\" (UID: \"24452fe2-b990-4537-b619-9db436ed5acb\") " pod="openstack/horizon-7c885c5cb8-46fsf" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.170350 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6968fbf848-w68ch"] Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.170712 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/24452fe2-b990-4537-b619-9db436ed5acb-logs\") pod \"horizon-7c885c5cb8-46fsf\" (UID: \"24452fe2-b990-4537-b619-9db436ed5acb\") " pod="openstack/horizon-7c885c5cb8-46fsf" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.172058 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/24452fe2-b990-4537-b619-9db436ed5acb-config-data\") pod \"horizon-7c885c5cb8-46fsf\" (UID: \"24452fe2-b990-4537-b619-9db436ed5acb\") " pod="openstack/horizon-7c885c5cb8-46fsf" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.174984 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/24452fe2-b990-4537-b619-9db436ed5acb-horizon-tls-certs\") pod \"horizon-7c885c5cb8-46fsf\" (UID: \"24452fe2-b990-4537-b619-9db436ed5acb\") " pod="openstack/horizon-7c885c5cb8-46fsf" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.175560 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/24452fe2-b990-4537-b619-9db436ed5acb-scripts\") pod \"horizon-7c885c5cb8-46fsf\" (UID: \"24452fe2-b990-4537-b619-9db436ed5acb\") " pod="openstack/horizon-7c885c5cb8-46fsf" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.179607 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24452fe2-b990-4537-b619-9db436ed5acb-combined-ca-bundle\") pod \"horizon-7c885c5cb8-46fsf\" (UID: \"24452fe2-b990-4537-b619-9db436ed5acb\") " pod="openstack/horizon-7c885c5cb8-46fsf" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.198535 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldm2p\" (UniqueName: \"kubernetes.io/projected/24452fe2-b990-4537-b619-9db436ed5acb-kube-api-access-ldm2p\") pod \"horizon-7c885c5cb8-46fsf\" (UID: \"24452fe2-b990-4537-b619-9db436ed5acb\") " pod="openstack/horizon-7c885c5cb8-46fsf" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.238265 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5dccf2f5-f149-4b29-99a5-560ff38df6f1","Type":"ContainerStarted","Data":"d786e841a947b0b189e97b8d78fb432ceb6f197185d2e0c481af1b81a19581b5"} Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.260817 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"354e9917-810a-44ba-bc85-b12e15152129","Type":"ContainerStarted","Data":"3dc60723b6966df88ba5c9c02f4bf0307620723e1f9b69392e836c0fe6c0ad8c"} Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.271698 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"96e55f05-de52-4914-9fc8-7634bff464a7","Type":"ContainerStarted","Data":"02e484f8e548edab79c84b048cc92d53ac7f7480e992a74097caa260cb6456aa"} Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.277879 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcn4f\" (UniqueName: \"kubernetes.io/projected/7f8ecf06-f050-454b-8d36-e91b49847601-kube-api-access-bcn4f\") pod \"horizon-6968fbf848-w68ch\" (UID: \"7f8ecf06-f050-454b-8d36-e91b49847601\") " pod="openstack/horizon-6968fbf848-w68ch" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.277964 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f8ecf06-f050-454b-8d36-e91b49847601-logs\") pod \"horizon-6968fbf848-w68ch\" (UID: \"7f8ecf06-f050-454b-8d36-e91b49847601\") " pod="openstack/horizon-6968fbf848-w68ch" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.277989 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f8ecf06-f050-454b-8d36-e91b49847601-scripts\") pod \"horizon-6968fbf848-w68ch\" (UID: \"7f8ecf06-f050-454b-8d36-e91b49847601\") " pod="openstack/horizon-6968fbf848-w68ch" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.278056 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7f8ecf06-f050-454b-8d36-e91b49847601-config-data\") pod \"horizon-6968fbf848-w68ch\" (UID: \"7f8ecf06-f050-454b-8d36-e91b49847601\") " pod="openstack/horizon-6968fbf848-w68ch" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.278109 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f8ecf06-f050-454b-8d36-e91b49847601-horizon-tls-certs\") pod \"horizon-6968fbf848-w68ch\" (UID: \"7f8ecf06-f050-454b-8d36-e91b49847601\") " pod="openstack/horizon-6968fbf848-w68ch" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.278158 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7f8ecf06-f050-454b-8d36-e91b49847601-horizon-secret-key\") pod \"horizon-6968fbf848-w68ch\" (UID: \"7f8ecf06-f050-454b-8d36-e91b49847601\") " pod="openstack/horizon-6968fbf848-w68ch" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.278205 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f8ecf06-f050-454b-8d36-e91b49847601-combined-ca-bundle\") pod \"horizon-6968fbf848-w68ch\" (UID: \"7f8ecf06-f050-454b-8d36-e91b49847601\") " pod="openstack/horizon-6968fbf848-w68ch" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.299628 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"3ba752bd-4b19-4ac4-9798-60de8b5a6f3e","Type":"ContainerStarted","Data":"a712040efb2f270d2f0615f4bf9379fccf98dc84821413f117dee91b3437479d"} Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.322366 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-volume1-0" podStartSLOduration=4.046611929 podStartE2EDuration="5.322341921s" podCreationTimestamp="2025-12-01 09:32:45 +0000 UTC" firstStartedPulling="2025-12-01 09:32:46.986471816 +0000 UTC m=+3142.888580365" lastFinishedPulling="2025-12-01 09:32:48.262201818 +0000 UTC m=+3144.164310357" observedRunningTime="2025-12-01 09:32:50.319579292 +0000 UTC m=+3146.221687831" watchObservedRunningTime="2025-12-01 09:32:50.322341921 +0000 UTC m=+3146.224450460" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.345643 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c885c5cb8-46fsf" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.372320 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=4.163508148 podStartE2EDuration="5.372303703s" podCreationTimestamp="2025-12-01 09:32:45 +0000 UTC" firstStartedPulling="2025-12-01 09:32:47.055376523 +0000 UTC m=+3142.957485062" lastFinishedPulling="2025-12-01 09:32:48.264172078 +0000 UTC m=+3144.166280617" observedRunningTime="2025-12-01 09:32:50.359348769 +0000 UTC m=+3146.261457308" watchObservedRunningTime="2025-12-01 09:32:50.372303703 +0000 UTC m=+3146.274412242" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.382296 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f8ecf06-f050-454b-8d36-e91b49847601-combined-ca-bundle\") pod \"horizon-6968fbf848-w68ch\" (UID: \"7f8ecf06-f050-454b-8d36-e91b49847601\") " pod="openstack/horizon-6968fbf848-w68ch" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.382748 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcn4f\" (UniqueName: \"kubernetes.io/projected/7f8ecf06-f050-454b-8d36-e91b49847601-kube-api-access-bcn4f\") pod \"horizon-6968fbf848-w68ch\" (UID: \"7f8ecf06-f050-454b-8d36-e91b49847601\") " pod="openstack/horizon-6968fbf848-w68ch" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.382921 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f8ecf06-f050-454b-8d36-e91b49847601-logs\") pod \"horizon-6968fbf848-w68ch\" (UID: \"7f8ecf06-f050-454b-8d36-e91b49847601\") " pod="openstack/horizon-6968fbf848-w68ch" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.382967 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f8ecf06-f050-454b-8d36-e91b49847601-scripts\") pod \"horizon-6968fbf848-w68ch\" (UID: \"7f8ecf06-f050-454b-8d36-e91b49847601\") " pod="openstack/horizon-6968fbf848-w68ch" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.383041 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7f8ecf06-f050-454b-8d36-e91b49847601-config-data\") pod \"horizon-6968fbf848-w68ch\" (UID: \"7f8ecf06-f050-454b-8d36-e91b49847601\") " pod="openstack/horizon-6968fbf848-w68ch" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.383142 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f8ecf06-f050-454b-8d36-e91b49847601-horizon-tls-certs\") pod \"horizon-6968fbf848-w68ch\" (UID: \"7f8ecf06-f050-454b-8d36-e91b49847601\") " pod="openstack/horizon-6968fbf848-w68ch" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.383235 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7f8ecf06-f050-454b-8d36-e91b49847601-horizon-secret-key\") pod \"horizon-6968fbf848-w68ch\" (UID: \"7f8ecf06-f050-454b-8d36-e91b49847601\") " pod="openstack/horizon-6968fbf848-w68ch" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.387946 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7f8ecf06-f050-454b-8d36-e91b49847601-horizon-secret-key\") pod \"horizon-6968fbf848-w68ch\" (UID: \"7f8ecf06-f050-454b-8d36-e91b49847601\") " pod="openstack/horizon-6968fbf848-w68ch" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.388741 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f8ecf06-f050-454b-8d36-e91b49847601-scripts\") pod \"horizon-6968fbf848-w68ch\" (UID: \"7f8ecf06-f050-454b-8d36-e91b49847601\") " pod="openstack/horizon-6968fbf848-w68ch" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.389461 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7f8ecf06-f050-454b-8d36-e91b49847601-config-data\") pod \"horizon-6968fbf848-w68ch\" (UID: \"7f8ecf06-f050-454b-8d36-e91b49847601\") " pod="openstack/horizon-6968fbf848-w68ch" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.391318 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f8ecf06-f050-454b-8d36-e91b49847601-combined-ca-bundle\") pod \"horizon-6968fbf848-w68ch\" (UID: \"7f8ecf06-f050-454b-8d36-e91b49847601\") " pod="openstack/horizon-6968fbf848-w68ch" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.393130 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f8ecf06-f050-454b-8d36-e91b49847601-logs\") pod \"horizon-6968fbf848-w68ch\" (UID: \"7f8ecf06-f050-454b-8d36-e91b49847601\") " pod="openstack/horizon-6968fbf848-w68ch" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.395246 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f8ecf06-f050-454b-8d36-e91b49847601-horizon-tls-certs\") pod \"horizon-6968fbf848-w68ch\" (UID: \"7f8ecf06-f050-454b-8d36-e91b49847601\") " pod="openstack/horizon-6968fbf848-w68ch" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.415347 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcn4f\" (UniqueName: \"kubernetes.io/projected/7f8ecf06-f050-454b-8d36-e91b49847601-kube-api-access-bcn4f\") pod \"horizon-6968fbf848-w68ch\" (UID: \"7f8ecf06-f050-454b-8d36-e91b49847601\") " pod="openstack/horizon-6968fbf848-w68ch" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.659752 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6968fbf848-w68ch" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.763000 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:50 crc kubenswrapper[4873]: I1201 09:32:50.819364 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Dec 01 09:32:51 crc kubenswrapper[4873]: I1201 09:32:51.079270 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7c885c5cb8-46fsf"] Dec 01 09:32:51 crc kubenswrapper[4873]: I1201 09:32:51.127507 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-9a52-account-create-update-vtg5d" Dec 01 09:32:51 crc kubenswrapper[4873]: W1201 09:32:51.139480 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod24452fe2_b990_4537_b619_9db436ed5acb.slice/crio-2398a1a8f1a9676fc632643ad733b0d8e0a617d5488c48bdcbe599ea953a68d1 WatchSource:0}: Error finding container 2398a1a8f1a9676fc632643ad733b0d8e0a617d5488c48bdcbe599ea953a68d1: Status 404 returned error can't find the container with id 2398a1a8f1a9676fc632643ad733b0d8e0a617d5488c48bdcbe599ea953a68d1 Dec 01 09:32:51 crc kubenswrapper[4873]: I1201 09:32:51.218756 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-v7b4d" Dec 01 09:32:51 crc kubenswrapper[4873]: I1201 09:32:51.240703 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmzqn\" (UniqueName: \"kubernetes.io/projected/483831f0-8b26-4845-b893-50bb7fb6ddb4-kube-api-access-vmzqn\") pod \"483831f0-8b26-4845-b893-50bb7fb6ddb4\" (UID: \"483831f0-8b26-4845-b893-50bb7fb6ddb4\") " Dec 01 09:32:51 crc kubenswrapper[4873]: I1201 09:32:51.241160 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/483831f0-8b26-4845-b893-50bb7fb6ddb4-operator-scripts\") pod \"483831f0-8b26-4845-b893-50bb7fb6ddb4\" (UID: \"483831f0-8b26-4845-b893-50bb7fb6ddb4\") " Dec 01 09:32:51 crc kubenswrapper[4873]: I1201 09:32:51.243525 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/483831f0-8b26-4845-b893-50bb7fb6ddb4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "483831f0-8b26-4845-b893-50bb7fb6ddb4" (UID: "483831f0-8b26-4845-b893-50bb7fb6ddb4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:32:51 crc kubenswrapper[4873]: I1201 09:32:51.254700 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/483831f0-8b26-4845-b893-50bb7fb6ddb4-kube-api-access-vmzqn" (OuterVolumeSpecName: "kube-api-access-vmzqn") pod "483831f0-8b26-4845-b893-50bb7fb6ddb4" (UID: "483831f0-8b26-4845-b893-50bb7fb6ddb4"). InnerVolumeSpecName "kube-api-access-vmzqn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:32:51 crc kubenswrapper[4873]: I1201 09:32:51.343882 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8a98544-560c-487a-9a71-59b53a882c9b-operator-scripts\") pod \"a8a98544-560c-487a-9a71-59b53a882c9b\" (UID: \"a8a98544-560c-487a-9a71-59b53a882c9b\") " Dec 01 09:32:51 crc kubenswrapper[4873]: I1201 09:32:51.344285 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8r9c5\" (UniqueName: \"kubernetes.io/projected/a8a98544-560c-487a-9a71-59b53a882c9b-kube-api-access-8r9c5\") pod \"a8a98544-560c-487a-9a71-59b53a882c9b\" (UID: \"a8a98544-560c-487a-9a71-59b53a882c9b\") " Dec 01 09:32:51 crc kubenswrapper[4873]: I1201 09:32:51.357758 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8a98544-560c-487a-9a71-59b53a882c9b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a8a98544-560c-487a-9a71-59b53a882c9b" (UID: "a8a98544-560c-487a-9a71-59b53a882c9b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:32:51 crc kubenswrapper[4873]: I1201 09:32:51.360646 4873 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8a98544-560c-487a-9a71-59b53a882c9b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:51 crc kubenswrapper[4873]: I1201 09:32:51.360716 4873 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/483831f0-8b26-4845-b893-50bb7fb6ddb4-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:51 crc kubenswrapper[4873]: I1201 09:32:51.360737 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmzqn\" (UniqueName: \"kubernetes.io/projected/483831f0-8b26-4845-b893-50bb7fb6ddb4-kube-api-access-vmzqn\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:51 crc kubenswrapper[4873]: I1201 09:32:51.364862 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8a98544-560c-487a-9a71-59b53a882c9b-kube-api-access-8r9c5" (OuterVolumeSpecName: "kube-api-access-8r9c5") pod "a8a98544-560c-487a-9a71-59b53a882c9b" (UID: "a8a98544-560c-487a-9a71-59b53a882c9b"). InnerVolumeSpecName "kube-api-access-8r9c5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:32:51 crc kubenswrapper[4873]: I1201 09:32:51.376271 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-9a52-account-create-update-vtg5d" Dec 01 09:32:51 crc kubenswrapper[4873]: I1201 09:32:51.376317 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-9a52-account-create-update-vtg5d" event={"ID":"483831f0-8b26-4845-b893-50bb7fb6ddb4","Type":"ContainerDied","Data":"fa31785d87239fe681ce10febf6e012eccd78ea7adf94b2285a244e3e27c8a6c"} Dec 01 09:32:51 crc kubenswrapper[4873]: I1201 09:32:51.376375 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa31785d87239fe681ce10febf6e012eccd78ea7adf94b2285a244e3e27c8a6c" Dec 01 09:32:51 crc kubenswrapper[4873]: I1201 09:32:51.384653 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c885c5cb8-46fsf" event={"ID":"24452fe2-b990-4537-b619-9db436ed5acb","Type":"ContainerStarted","Data":"2398a1a8f1a9676fc632643ad733b0d8e0a617d5488c48bdcbe599ea953a68d1"} Dec 01 09:32:51 crc kubenswrapper[4873]: I1201 09:32:51.394093 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-v7b4d" Dec 01 09:32:51 crc kubenswrapper[4873]: I1201 09:32:51.394187 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-v7b4d" event={"ID":"a8a98544-560c-487a-9a71-59b53a882c9b","Type":"ContainerDied","Data":"492c39da1084067e6b717eb2043de864da8215bd57280dd57d132a0499eb4b88"} Dec 01 09:32:51 crc kubenswrapper[4873]: I1201 09:32:51.394267 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="492c39da1084067e6b717eb2043de864da8215bd57280dd57d132a0499eb4b88" Dec 01 09:32:51 crc kubenswrapper[4873]: I1201 09:32:51.455626 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6968fbf848-w68ch"] Dec 01 09:32:51 crc kubenswrapper[4873]: I1201 09:32:51.467758 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8r9c5\" (UniqueName: \"kubernetes.io/projected/a8a98544-560c-487a-9a71-59b53a882c9b-kube-api-access-8r9c5\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:52 crc kubenswrapper[4873]: I1201 09:32:52.413727 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"354e9917-810a-44ba-bc85-b12e15152129","Type":"ContainerStarted","Data":"83d93d861e3da39c9d35287c95f638f205649705dd38c2a91f271910767f2959"} Dec 01 09:32:52 crc kubenswrapper[4873]: I1201 09:32:52.414505 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"354e9917-810a-44ba-bc85-b12e15152129","Type":"ContainerStarted","Data":"169d88b15c9af03c1f05649721e18974247cac0100079ecf92d681923f66c705"} Dec 01 09:32:52 crc kubenswrapper[4873]: I1201 09:32:52.413952 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="354e9917-810a-44ba-bc85-b12e15152129" containerName="glance-httpd" containerID="cri-o://169d88b15c9af03c1f05649721e18974247cac0100079ecf92d681923f66c705" gracePeriod=30 Dec 01 09:32:52 crc kubenswrapper[4873]: I1201 09:32:52.413938 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="354e9917-810a-44ba-bc85-b12e15152129" containerName="glance-log" containerID="cri-o://83d93d861e3da39c9d35287c95f638f205649705dd38c2a91f271910767f2959" gracePeriod=30 Dec 01 09:32:52 crc kubenswrapper[4873]: I1201 09:32:52.420388 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6968fbf848-w68ch" event={"ID":"7f8ecf06-f050-454b-8d36-e91b49847601","Type":"ContainerStarted","Data":"314bebab88d370147ead9dc45e0351abb615e211fa91460c2a71eed15ae0b721"} Dec 01 09:32:52 crc kubenswrapper[4873]: I1201 09:32:52.429738 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5dccf2f5-f149-4b29-99a5-560ff38df6f1" containerName="glance-log" containerID="cri-o://eee8f5a9d20f79a06e4d66c9ee91eb14e0e738dc0b51278702a6db299c5e7d7c" gracePeriod=30 Dec 01 09:32:52 crc kubenswrapper[4873]: I1201 09:32:52.429899 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5dccf2f5-f149-4b29-99a5-560ff38df6f1" containerName="glance-httpd" containerID="cri-o://c54d86d6fcf7ffbc2e1dcc773c6aca1c40e3b8647426a757f8269f6eca5020e0" gracePeriod=30 Dec 01 09:32:52 crc kubenswrapper[4873]: I1201 09:32:52.429764 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5dccf2f5-f149-4b29-99a5-560ff38df6f1","Type":"ContainerStarted","Data":"eee8f5a9d20f79a06e4d66c9ee91eb14e0e738dc0b51278702a6db299c5e7d7c"} Dec 01 09:32:52 crc kubenswrapper[4873]: I1201 09:32:52.430703 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5dccf2f5-f149-4b29-99a5-560ff38df6f1","Type":"ContainerStarted","Data":"c54d86d6fcf7ffbc2e1dcc773c6aca1c40e3b8647426a757f8269f6eca5020e0"} Dec 01 09:32:52 crc kubenswrapper[4873]: I1201 09:32:52.477253 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.477229848 podStartE2EDuration="4.477229848s" podCreationTimestamp="2025-12-01 09:32:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:32:52.446335414 +0000 UTC m=+3148.348443963" watchObservedRunningTime="2025-12-01 09:32:52.477229848 +0000 UTC m=+3148.379338387" Dec 01 09:32:52 crc kubenswrapper[4873]: I1201 09:32:52.486986 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.486959822 podStartE2EDuration="4.486959822s" podCreationTimestamp="2025-12-01 09:32:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:32:52.475419893 +0000 UTC m=+3148.377528432" watchObservedRunningTime="2025-12-01 09:32:52.486959822 +0000 UTC m=+3148.389068361" Dec 01 09:32:52 crc kubenswrapper[4873]: E1201 09:32:52.782782 4873 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5dccf2f5_f149_4b29_99a5_560ff38df6f1.slice/crio-eee8f5a9d20f79a06e4d66c9ee91eb14e0e738dc0b51278702a6db299c5e7d7c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5dccf2f5_f149_4b29_99a5_560ff38df6f1.slice/crio-conmon-eee8f5a9d20f79a06e4d66c9ee91eb14e0e738dc0b51278702a6db299c5e7d7c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5dccf2f5_f149_4b29_99a5_560ff38df6f1.slice/crio-c54d86d6fcf7ffbc2e1dcc773c6aca1c40e3b8647426a757f8269f6eca5020e0.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5dccf2f5_f149_4b29_99a5_560ff38df6f1.slice/crio-conmon-c54d86d6fcf7ffbc2e1dcc773c6aca1c40e3b8647426a757f8269f6eca5020e0.scope\": RecentStats: unable to find data in memory cache]" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.122335 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.224906 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/354e9917-810a-44ba-bc85-b12e15152129-combined-ca-bundle\") pod \"354e9917-810a-44ba-bc85-b12e15152129\" (UID: \"354e9917-810a-44ba-bc85-b12e15152129\") " Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.225001 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/354e9917-810a-44ba-bc85-b12e15152129-logs\") pod \"354e9917-810a-44ba-bc85-b12e15152129\" (UID: \"354e9917-810a-44ba-bc85-b12e15152129\") " Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.225047 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"354e9917-810a-44ba-bc85-b12e15152129\" (UID: \"354e9917-810a-44ba-bc85-b12e15152129\") " Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.225132 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/354e9917-810a-44ba-bc85-b12e15152129-ceph\") pod \"354e9917-810a-44ba-bc85-b12e15152129\" (UID: \"354e9917-810a-44ba-bc85-b12e15152129\") " Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.225167 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/354e9917-810a-44ba-bc85-b12e15152129-scripts\") pod \"354e9917-810a-44ba-bc85-b12e15152129\" (UID: \"354e9917-810a-44ba-bc85-b12e15152129\") " Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.225226 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/354e9917-810a-44ba-bc85-b12e15152129-httpd-run\") pod \"354e9917-810a-44ba-bc85-b12e15152129\" (UID: \"354e9917-810a-44ba-bc85-b12e15152129\") " Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.225400 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rkmtw\" (UniqueName: \"kubernetes.io/projected/354e9917-810a-44ba-bc85-b12e15152129-kube-api-access-rkmtw\") pod \"354e9917-810a-44ba-bc85-b12e15152129\" (UID: \"354e9917-810a-44ba-bc85-b12e15152129\") " Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.225458 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/354e9917-810a-44ba-bc85-b12e15152129-config-data\") pod \"354e9917-810a-44ba-bc85-b12e15152129\" (UID: \"354e9917-810a-44ba-bc85-b12e15152129\") " Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.225537 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/354e9917-810a-44ba-bc85-b12e15152129-public-tls-certs\") pod \"354e9917-810a-44ba-bc85-b12e15152129\" (UID: \"354e9917-810a-44ba-bc85-b12e15152129\") " Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.226289 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/354e9917-810a-44ba-bc85-b12e15152129-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "354e9917-810a-44ba-bc85-b12e15152129" (UID: "354e9917-810a-44ba-bc85-b12e15152129"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.226443 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/354e9917-810a-44ba-bc85-b12e15152129-logs" (OuterVolumeSpecName: "logs") pod "354e9917-810a-44ba-bc85-b12e15152129" (UID: "354e9917-810a-44ba-bc85-b12e15152129"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.227351 4873 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/354e9917-810a-44ba-bc85-b12e15152129-logs\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.227375 4873 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/354e9917-810a-44ba-bc85-b12e15152129-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.232287 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/354e9917-810a-44ba-bc85-b12e15152129-scripts" (OuterVolumeSpecName: "scripts") pod "354e9917-810a-44ba-bc85-b12e15152129" (UID: "354e9917-810a-44ba-bc85-b12e15152129"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.233740 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "354e9917-810a-44ba-bc85-b12e15152129" (UID: "354e9917-810a-44ba-bc85-b12e15152129"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.235693 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/354e9917-810a-44ba-bc85-b12e15152129-ceph" (OuterVolumeSpecName: "ceph") pod "354e9917-810a-44ba-bc85-b12e15152129" (UID: "354e9917-810a-44ba-bc85-b12e15152129"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.248554 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/354e9917-810a-44ba-bc85-b12e15152129-kube-api-access-rkmtw" (OuterVolumeSpecName: "kube-api-access-rkmtw") pod "354e9917-810a-44ba-bc85-b12e15152129" (UID: "354e9917-810a-44ba-bc85-b12e15152129"). InnerVolumeSpecName "kube-api-access-rkmtw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.278529 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/354e9917-810a-44ba-bc85-b12e15152129-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "354e9917-810a-44ba-bc85-b12e15152129" (UID: "354e9917-810a-44ba-bc85-b12e15152129"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.286581 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/354e9917-810a-44ba-bc85-b12e15152129-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "354e9917-810a-44ba-bc85-b12e15152129" (UID: "354e9917-810a-44ba-bc85-b12e15152129"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.290061 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/354e9917-810a-44ba-bc85-b12e15152129-config-data" (OuterVolumeSpecName: "config-data") pod "354e9917-810a-44ba-bc85-b12e15152129" (UID: "354e9917-810a-44ba-bc85-b12e15152129"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.325976 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.330292 4873 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/354e9917-810a-44ba-bc85-b12e15152129-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.330327 4873 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/354e9917-810a-44ba-bc85-b12e15152129-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.330359 4873 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.330371 4873 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/354e9917-810a-44ba-bc85-b12e15152129-ceph\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.330382 4873 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/354e9917-810a-44ba-bc85-b12e15152129-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.330392 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rkmtw\" (UniqueName: \"kubernetes.io/projected/354e9917-810a-44ba-bc85-b12e15152129-kube-api-access-rkmtw\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.330402 4873 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/354e9917-810a-44ba-bc85-b12e15152129-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.355545 4873 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.431752 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5dccf2f5-f149-4b29-99a5-560ff38df6f1-scripts\") pod \"5dccf2f5-f149-4b29-99a5-560ff38df6f1\" (UID: \"5dccf2f5-f149-4b29-99a5-560ff38df6f1\") " Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.431999 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htdzz\" (UniqueName: \"kubernetes.io/projected/5dccf2f5-f149-4b29-99a5-560ff38df6f1-kube-api-access-htdzz\") pod \"5dccf2f5-f149-4b29-99a5-560ff38df6f1\" (UID: \"5dccf2f5-f149-4b29-99a5-560ff38df6f1\") " Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.432077 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"5dccf2f5-f149-4b29-99a5-560ff38df6f1\" (UID: \"5dccf2f5-f149-4b29-99a5-560ff38df6f1\") " Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.432152 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5dccf2f5-f149-4b29-99a5-560ff38df6f1-httpd-run\") pod \"5dccf2f5-f149-4b29-99a5-560ff38df6f1\" (UID: \"5dccf2f5-f149-4b29-99a5-560ff38df6f1\") " Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.432284 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5dccf2f5-f149-4b29-99a5-560ff38df6f1-ceph\") pod \"5dccf2f5-f149-4b29-99a5-560ff38df6f1\" (UID: \"5dccf2f5-f149-4b29-99a5-560ff38df6f1\") " Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.432346 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5dccf2f5-f149-4b29-99a5-560ff38df6f1-combined-ca-bundle\") pod \"5dccf2f5-f149-4b29-99a5-560ff38df6f1\" (UID: \"5dccf2f5-f149-4b29-99a5-560ff38df6f1\") " Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.432376 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5dccf2f5-f149-4b29-99a5-560ff38df6f1-logs\") pod \"5dccf2f5-f149-4b29-99a5-560ff38df6f1\" (UID: \"5dccf2f5-f149-4b29-99a5-560ff38df6f1\") " Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.432432 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5dccf2f5-f149-4b29-99a5-560ff38df6f1-internal-tls-certs\") pod \"5dccf2f5-f149-4b29-99a5-560ff38df6f1\" (UID: \"5dccf2f5-f149-4b29-99a5-560ff38df6f1\") " Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.432476 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5dccf2f5-f149-4b29-99a5-560ff38df6f1-config-data\") pod \"5dccf2f5-f149-4b29-99a5-560ff38df6f1\" (UID: \"5dccf2f5-f149-4b29-99a5-560ff38df6f1\") " Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.432676 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5dccf2f5-f149-4b29-99a5-560ff38df6f1-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "5dccf2f5-f149-4b29-99a5-560ff38df6f1" (UID: "5dccf2f5-f149-4b29-99a5-560ff38df6f1"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.432818 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5dccf2f5-f149-4b29-99a5-560ff38df6f1-logs" (OuterVolumeSpecName: "logs") pod "5dccf2f5-f149-4b29-99a5-560ff38df6f1" (UID: "5dccf2f5-f149-4b29-99a5-560ff38df6f1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.433672 4873 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5dccf2f5-f149-4b29-99a5-560ff38df6f1-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.433730 4873 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.433755 4873 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5dccf2f5-f149-4b29-99a5-560ff38df6f1-logs\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.436604 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "5dccf2f5-f149-4b29-99a5-560ff38df6f1" (UID: "5dccf2f5-f149-4b29-99a5-560ff38df6f1"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.437665 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5dccf2f5-f149-4b29-99a5-560ff38df6f1-kube-api-access-htdzz" (OuterVolumeSpecName: "kube-api-access-htdzz") pod "5dccf2f5-f149-4b29-99a5-560ff38df6f1" (UID: "5dccf2f5-f149-4b29-99a5-560ff38df6f1"). InnerVolumeSpecName "kube-api-access-htdzz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.437822 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5dccf2f5-f149-4b29-99a5-560ff38df6f1-scripts" (OuterVolumeSpecName: "scripts") pod "5dccf2f5-f149-4b29-99a5-560ff38df6f1" (UID: "5dccf2f5-f149-4b29-99a5-560ff38df6f1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.439436 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5dccf2f5-f149-4b29-99a5-560ff38df6f1-ceph" (OuterVolumeSpecName: "ceph") pod "5dccf2f5-f149-4b29-99a5-560ff38df6f1" (UID: "5dccf2f5-f149-4b29-99a5-560ff38df6f1"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.450458 4873 generic.go:334] "Generic (PLEG): container finished" podID="5dccf2f5-f149-4b29-99a5-560ff38df6f1" containerID="c54d86d6fcf7ffbc2e1dcc773c6aca1c40e3b8647426a757f8269f6eca5020e0" exitCode=143 Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.450516 4873 generic.go:334] "Generic (PLEG): container finished" podID="5dccf2f5-f149-4b29-99a5-560ff38df6f1" containerID="eee8f5a9d20f79a06e4d66c9ee91eb14e0e738dc0b51278702a6db299c5e7d7c" exitCode=143 Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.450540 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.450542 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5dccf2f5-f149-4b29-99a5-560ff38df6f1","Type":"ContainerDied","Data":"c54d86d6fcf7ffbc2e1dcc773c6aca1c40e3b8647426a757f8269f6eca5020e0"} Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.450631 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5dccf2f5-f149-4b29-99a5-560ff38df6f1","Type":"ContainerDied","Data":"eee8f5a9d20f79a06e4d66c9ee91eb14e0e738dc0b51278702a6db299c5e7d7c"} Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.450645 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5dccf2f5-f149-4b29-99a5-560ff38df6f1","Type":"ContainerDied","Data":"d786e841a947b0b189e97b8d78fb432ceb6f197185d2e0c481af1b81a19581b5"} Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.450674 4873 scope.go:117] "RemoveContainer" containerID="c54d86d6fcf7ffbc2e1dcc773c6aca1c40e3b8647426a757f8269f6eca5020e0" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.456988 4873 generic.go:334] "Generic (PLEG): container finished" podID="354e9917-810a-44ba-bc85-b12e15152129" containerID="169d88b15c9af03c1f05649721e18974247cac0100079ecf92d681923f66c705" exitCode=143 Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.457079 4873 generic.go:334] "Generic (PLEG): container finished" podID="354e9917-810a-44ba-bc85-b12e15152129" containerID="83d93d861e3da39c9d35287c95f638f205649705dd38c2a91f271910767f2959" exitCode=143 Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.457061 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.457062 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"354e9917-810a-44ba-bc85-b12e15152129","Type":"ContainerDied","Data":"169d88b15c9af03c1f05649721e18974247cac0100079ecf92d681923f66c705"} Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.457180 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"354e9917-810a-44ba-bc85-b12e15152129","Type":"ContainerDied","Data":"83d93d861e3da39c9d35287c95f638f205649705dd38c2a91f271910767f2959"} Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.457219 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"354e9917-810a-44ba-bc85-b12e15152129","Type":"ContainerDied","Data":"3dc60723b6966df88ba5c9c02f4bf0307620723e1f9b69392e836c0fe6c0ad8c"} Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.468791 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5dccf2f5-f149-4b29-99a5-560ff38df6f1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5dccf2f5-f149-4b29-99a5-560ff38df6f1" (UID: "5dccf2f5-f149-4b29-99a5-560ff38df6f1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.498867 4873 scope.go:117] "RemoveContainer" containerID="eee8f5a9d20f79a06e4d66c9ee91eb14e0e738dc0b51278702a6db299c5e7d7c" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.509069 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.511598 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5dccf2f5-f149-4b29-99a5-560ff38df6f1-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "5dccf2f5-f149-4b29-99a5-560ff38df6f1" (UID: "5dccf2f5-f149-4b29-99a5-560ff38df6f1"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.511668 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5dccf2f5-f149-4b29-99a5-560ff38df6f1-config-data" (OuterVolumeSpecName: "config-data") pod "5dccf2f5-f149-4b29-99a5-560ff38df6f1" (UID: "5dccf2f5-f149-4b29-99a5-560ff38df6f1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.523961 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.543298 4873 scope.go:117] "RemoveContainer" containerID="c54d86d6fcf7ffbc2e1dcc773c6aca1c40e3b8647426a757f8269f6eca5020e0" Dec 01 09:32:53 crc kubenswrapper[4873]: E1201 09:32:53.543896 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c54d86d6fcf7ffbc2e1dcc773c6aca1c40e3b8647426a757f8269f6eca5020e0\": container with ID starting with c54d86d6fcf7ffbc2e1dcc773c6aca1c40e3b8647426a757f8269f6eca5020e0 not found: ID does not exist" containerID="c54d86d6fcf7ffbc2e1dcc773c6aca1c40e3b8647426a757f8269f6eca5020e0" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.543927 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c54d86d6fcf7ffbc2e1dcc773c6aca1c40e3b8647426a757f8269f6eca5020e0"} err="failed to get container status \"c54d86d6fcf7ffbc2e1dcc773c6aca1c40e3b8647426a757f8269f6eca5020e0\": rpc error: code = NotFound desc = could not find container \"c54d86d6fcf7ffbc2e1dcc773c6aca1c40e3b8647426a757f8269f6eca5020e0\": container with ID starting with c54d86d6fcf7ffbc2e1dcc773c6aca1c40e3b8647426a757f8269f6eca5020e0 not found: ID does not exist" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.543955 4873 scope.go:117] "RemoveContainer" containerID="eee8f5a9d20f79a06e4d66c9ee91eb14e0e738dc0b51278702a6db299c5e7d7c" Dec 01 09:32:53 crc kubenswrapper[4873]: E1201 09:32:53.545695 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eee8f5a9d20f79a06e4d66c9ee91eb14e0e738dc0b51278702a6db299c5e7d7c\": container with ID starting with eee8f5a9d20f79a06e4d66c9ee91eb14e0e738dc0b51278702a6db299c5e7d7c not found: ID does not exist" containerID="eee8f5a9d20f79a06e4d66c9ee91eb14e0e738dc0b51278702a6db299c5e7d7c" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.545735 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eee8f5a9d20f79a06e4d66c9ee91eb14e0e738dc0b51278702a6db299c5e7d7c"} err="failed to get container status \"eee8f5a9d20f79a06e4d66c9ee91eb14e0e738dc0b51278702a6db299c5e7d7c\": rpc error: code = NotFound desc = could not find container \"eee8f5a9d20f79a06e4d66c9ee91eb14e0e738dc0b51278702a6db299c5e7d7c\": container with ID starting with eee8f5a9d20f79a06e4d66c9ee91eb14e0e738dc0b51278702a6db299c5e7d7c not found: ID does not exist" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.545755 4873 scope.go:117] "RemoveContainer" containerID="c54d86d6fcf7ffbc2e1dcc773c6aca1c40e3b8647426a757f8269f6eca5020e0" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.545866 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htdzz\" (UniqueName: \"kubernetes.io/projected/5dccf2f5-f149-4b29-99a5-560ff38df6f1-kube-api-access-htdzz\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.545919 4873 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.545933 4873 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5dccf2f5-f149-4b29-99a5-560ff38df6f1-ceph\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.545944 4873 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5dccf2f5-f149-4b29-99a5-560ff38df6f1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.545954 4873 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5dccf2f5-f149-4b29-99a5-560ff38df6f1-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.545963 4873 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5dccf2f5-f149-4b29-99a5-560ff38df6f1-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.545973 4873 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5dccf2f5-f149-4b29-99a5-560ff38df6f1-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.547989 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c54d86d6fcf7ffbc2e1dcc773c6aca1c40e3b8647426a757f8269f6eca5020e0"} err="failed to get container status \"c54d86d6fcf7ffbc2e1dcc773c6aca1c40e3b8647426a757f8269f6eca5020e0\": rpc error: code = NotFound desc = could not find container \"c54d86d6fcf7ffbc2e1dcc773c6aca1c40e3b8647426a757f8269f6eca5020e0\": container with ID starting with c54d86d6fcf7ffbc2e1dcc773c6aca1c40e3b8647426a757f8269f6eca5020e0 not found: ID does not exist" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.548047 4873 scope.go:117] "RemoveContainer" containerID="eee8f5a9d20f79a06e4d66c9ee91eb14e0e738dc0b51278702a6db299c5e7d7c" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.549101 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eee8f5a9d20f79a06e4d66c9ee91eb14e0e738dc0b51278702a6db299c5e7d7c"} err="failed to get container status \"eee8f5a9d20f79a06e4d66c9ee91eb14e0e738dc0b51278702a6db299c5e7d7c\": rpc error: code = NotFound desc = could not find container \"eee8f5a9d20f79a06e4d66c9ee91eb14e0e738dc0b51278702a6db299c5e7d7c\": container with ID starting with eee8f5a9d20f79a06e4d66c9ee91eb14e0e738dc0b51278702a6db299c5e7d7c not found: ID does not exist" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.549119 4873 scope.go:117] "RemoveContainer" containerID="169d88b15c9af03c1f05649721e18974247cac0100079ecf92d681923f66c705" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.553064 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 09:32:53 crc kubenswrapper[4873]: E1201 09:32:53.553776 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="354e9917-810a-44ba-bc85-b12e15152129" containerName="glance-httpd" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.554000 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="354e9917-810a-44ba-bc85-b12e15152129" containerName="glance-httpd" Dec 01 09:32:53 crc kubenswrapper[4873]: E1201 09:32:53.554202 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="483831f0-8b26-4845-b893-50bb7fb6ddb4" containerName="mariadb-account-create-update" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.554290 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="483831f0-8b26-4845-b893-50bb7fb6ddb4" containerName="mariadb-account-create-update" Dec 01 09:32:53 crc kubenswrapper[4873]: E1201 09:32:53.554385 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="354e9917-810a-44ba-bc85-b12e15152129" containerName="glance-log" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.554480 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="354e9917-810a-44ba-bc85-b12e15152129" containerName="glance-log" Dec 01 09:32:53 crc kubenswrapper[4873]: E1201 09:32:53.554662 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dccf2f5-f149-4b29-99a5-560ff38df6f1" containerName="glance-httpd" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.554766 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dccf2f5-f149-4b29-99a5-560ff38df6f1" containerName="glance-httpd" Dec 01 09:32:53 crc kubenswrapper[4873]: E1201 09:32:53.554886 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8a98544-560c-487a-9a71-59b53a882c9b" containerName="mariadb-database-create" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.554982 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8a98544-560c-487a-9a71-59b53a882c9b" containerName="mariadb-database-create" Dec 01 09:32:53 crc kubenswrapper[4873]: E1201 09:32:53.555135 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dccf2f5-f149-4b29-99a5-560ff38df6f1" containerName="glance-log" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.555276 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dccf2f5-f149-4b29-99a5-560ff38df6f1" containerName="glance-log" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.555737 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="354e9917-810a-44ba-bc85-b12e15152129" containerName="glance-httpd" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.555840 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="483831f0-8b26-4845-b893-50bb7fb6ddb4" containerName="mariadb-account-create-update" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.555960 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="5dccf2f5-f149-4b29-99a5-560ff38df6f1" containerName="glance-httpd" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.556085 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="354e9917-810a-44ba-bc85-b12e15152129" containerName="glance-log" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.556161 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8a98544-560c-487a-9a71-59b53a882c9b" containerName="mariadb-database-create" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.556227 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="5dccf2f5-f149-4b29-99a5-560ff38df6f1" containerName="glance-log" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.558499 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.575127 4873 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.575837 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.576968 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.635534 4873 scope.go:117] "RemoveContainer" containerID="83d93d861e3da39c9d35287c95f638f205649705dd38c2a91f271910767f2959" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.637028 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.653895 4873 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.675030 4873 scope.go:117] "RemoveContainer" containerID="169d88b15c9af03c1f05649721e18974247cac0100079ecf92d681923f66c705" Dec 01 09:32:53 crc kubenswrapper[4873]: E1201 09:32:53.675934 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"169d88b15c9af03c1f05649721e18974247cac0100079ecf92d681923f66c705\": container with ID starting with 169d88b15c9af03c1f05649721e18974247cac0100079ecf92d681923f66c705 not found: ID does not exist" containerID="169d88b15c9af03c1f05649721e18974247cac0100079ecf92d681923f66c705" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.676008 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"169d88b15c9af03c1f05649721e18974247cac0100079ecf92d681923f66c705"} err="failed to get container status \"169d88b15c9af03c1f05649721e18974247cac0100079ecf92d681923f66c705\": rpc error: code = NotFound desc = could not find container \"169d88b15c9af03c1f05649721e18974247cac0100079ecf92d681923f66c705\": container with ID starting with 169d88b15c9af03c1f05649721e18974247cac0100079ecf92d681923f66c705 not found: ID does not exist" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.676151 4873 scope.go:117] "RemoveContainer" containerID="83d93d861e3da39c9d35287c95f638f205649705dd38c2a91f271910767f2959" Dec 01 09:32:53 crc kubenswrapper[4873]: E1201 09:32:53.677195 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83d93d861e3da39c9d35287c95f638f205649705dd38c2a91f271910767f2959\": container with ID starting with 83d93d861e3da39c9d35287c95f638f205649705dd38c2a91f271910767f2959 not found: ID does not exist" containerID="83d93d861e3da39c9d35287c95f638f205649705dd38c2a91f271910767f2959" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.677373 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83d93d861e3da39c9d35287c95f638f205649705dd38c2a91f271910767f2959"} err="failed to get container status \"83d93d861e3da39c9d35287c95f638f205649705dd38c2a91f271910767f2959\": rpc error: code = NotFound desc = could not find container \"83d93d861e3da39c9d35287c95f638f205649705dd38c2a91f271910767f2959\": container with ID starting with 83d93d861e3da39c9d35287c95f638f205649705dd38c2a91f271910767f2959 not found: ID does not exist" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.677544 4873 scope.go:117] "RemoveContainer" containerID="169d88b15c9af03c1f05649721e18974247cac0100079ecf92d681923f66c705" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.682701 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"169d88b15c9af03c1f05649721e18974247cac0100079ecf92d681923f66c705"} err="failed to get container status \"169d88b15c9af03c1f05649721e18974247cac0100079ecf92d681923f66c705\": rpc error: code = NotFound desc = could not find container \"169d88b15c9af03c1f05649721e18974247cac0100079ecf92d681923f66c705\": container with ID starting with 169d88b15c9af03c1f05649721e18974247cac0100079ecf92d681923f66c705 not found: ID does not exist" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.682746 4873 scope.go:117] "RemoveContainer" containerID="83d93d861e3da39c9d35287c95f638f205649705dd38c2a91f271910767f2959" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.683382 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83d93d861e3da39c9d35287c95f638f205649705dd38c2a91f271910767f2959"} err="failed to get container status \"83d93d861e3da39c9d35287c95f638f205649705dd38c2a91f271910767f2959\": rpc error: code = NotFound desc = could not find container \"83d93d861e3da39c9d35287c95f638f205649705dd38c2a91f271910767f2959\": container with ID starting with 83d93d861e3da39c9d35287c95f638f205649705dd38c2a91f271910767f2959 not found: ID does not exist" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.755989 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/380e16b7-2a33-4591-bb99-9382512be92c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"380e16b7-2a33-4591-bb99-9382512be92c\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.756075 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/380e16b7-2a33-4591-bb99-9382512be92c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"380e16b7-2a33-4591-bb99-9382512be92c\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.756138 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/380e16b7-2a33-4591-bb99-9382512be92c-logs\") pod \"glance-default-external-api-0\" (UID: \"380e16b7-2a33-4591-bb99-9382512be92c\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.756161 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/380e16b7-2a33-4591-bb99-9382512be92c-ceph\") pod \"glance-default-external-api-0\" (UID: \"380e16b7-2a33-4591-bb99-9382512be92c\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.756186 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/380e16b7-2a33-4591-bb99-9382512be92c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"380e16b7-2a33-4591-bb99-9382512be92c\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.756252 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/380e16b7-2a33-4591-bb99-9382512be92c-scripts\") pod \"glance-default-external-api-0\" (UID: \"380e16b7-2a33-4591-bb99-9382512be92c\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.756298 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/380e16b7-2a33-4591-bb99-9382512be92c-config-data\") pod \"glance-default-external-api-0\" (UID: \"380e16b7-2a33-4591-bb99-9382512be92c\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.756317 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6gxb\" (UniqueName: \"kubernetes.io/projected/380e16b7-2a33-4591-bb99-9382512be92c-kube-api-access-k6gxb\") pod \"glance-default-external-api-0\" (UID: \"380e16b7-2a33-4591-bb99-9382512be92c\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.756338 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"380e16b7-2a33-4591-bb99-9382512be92c\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.801052 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.812693 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.833914 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.836373 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.842553 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.843603 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.852908 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.858650 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/380e16b7-2a33-4591-bb99-9382512be92c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"380e16b7-2a33-4591-bb99-9382512be92c\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.858746 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/380e16b7-2a33-4591-bb99-9382512be92c-logs\") pod \"glance-default-external-api-0\" (UID: \"380e16b7-2a33-4591-bb99-9382512be92c\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.858806 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/380e16b7-2a33-4591-bb99-9382512be92c-ceph\") pod \"glance-default-external-api-0\" (UID: \"380e16b7-2a33-4591-bb99-9382512be92c\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.858834 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/380e16b7-2a33-4591-bb99-9382512be92c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"380e16b7-2a33-4591-bb99-9382512be92c\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.858913 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/380e16b7-2a33-4591-bb99-9382512be92c-scripts\") pod \"glance-default-external-api-0\" (UID: \"380e16b7-2a33-4591-bb99-9382512be92c\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.858959 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/380e16b7-2a33-4591-bb99-9382512be92c-config-data\") pod \"glance-default-external-api-0\" (UID: \"380e16b7-2a33-4591-bb99-9382512be92c\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.858979 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6gxb\" (UniqueName: \"kubernetes.io/projected/380e16b7-2a33-4591-bb99-9382512be92c-kube-api-access-k6gxb\") pod \"glance-default-external-api-0\" (UID: \"380e16b7-2a33-4591-bb99-9382512be92c\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.858998 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"380e16b7-2a33-4591-bb99-9382512be92c\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.859085 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/380e16b7-2a33-4591-bb99-9382512be92c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"380e16b7-2a33-4591-bb99-9382512be92c\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.859481 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/380e16b7-2a33-4591-bb99-9382512be92c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"380e16b7-2a33-4591-bb99-9382512be92c\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.860910 4873 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"380e16b7-2a33-4591-bb99-9382512be92c\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-external-api-0" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.861186 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/380e16b7-2a33-4591-bb99-9382512be92c-logs\") pod \"glance-default-external-api-0\" (UID: \"380e16b7-2a33-4591-bb99-9382512be92c\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.863121 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/380e16b7-2a33-4591-bb99-9382512be92c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"380e16b7-2a33-4591-bb99-9382512be92c\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.867060 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/380e16b7-2a33-4591-bb99-9382512be92c-scripts\") pod \"glance-default-external-api-0\" (UID: \"380e16b7-2a33-4591-bb99-9382512be92c\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.878854 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/380e16b7-2a33-4591-bb99-9382512be92c-config-data\") pod \"glance-default-external-api-0\" (UID: \"380e16b7-2a33-4591-bb99-9382512be92c\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.882860 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6gxb\" (UniqueName: \"kubernetes.io/projected/380e16b7-2a33-4591-bb99-9382512be92c-kube-api-access-k6gxb\") pod \"glance-default-external-api-0\" (UID: \"380e16b7-2a33-4591-bb99-9382512be92c\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.895741 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/380e16b7-2a33-4591-bb99-9382512be92c-ceph\") pod \"glance-default-external-api-0\" (UID: \"380e16b7-2a33-4591-bb99-9382512be92c\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.911846 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/380e16b7-2a33-4591-bb99-9382512be92c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"380e16b7-2a33-4591-bb99-9382512be92c\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.938812 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"380e16b7-2a33-4591-bb99-9382512be92c\") " pod="openstack/glance-default-external-api-0" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.961838 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90a86592-ced2-4308-91fc-e9993667556a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"90a86592-ced2-4308-91fc-e9993667556a\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.961936 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90a86592-ced2-4308-91fc-e9993667556a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"90a86592-ced2-4308-91fc-e9993667556a\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.961959 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90a86592-ced2-4308-91fc-e9993667556a-logs\") pod \"glance-default-internal-api-0\" (UID: \"90a86592-ced2-4308-91fc-e9993667556a\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.962102 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90a86592-ced2-4308-91fc-e9993667556a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"90a86592-ced2-4308-91fc-e9993667556a\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.962143 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/90a86592-ced2-4308-91fc-e9993667556a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"90a86592-ced2-4308-91fc-e9993667556a\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.962184 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4blg\" (UniqueName: \"kubernetes.io/projected/90a86592-ced2-4308-91fc-e9993667556a-kube-api-access-z4blg\") pod \"glance-default-internal-api-0\" (UID: \"90a86592-ced2-4308-91fc-e9993667556a\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.962211 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"90a86592-ced2-4308-91fc-e9993667556a\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.962266 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/90a86592-ced2-4308-91fc-e9993667556a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"90a86592-ced2-4308-91fc-e9993667556a\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:53 crc kubenswrapper[4873]: I1201 09:32:53.962294 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/90a86592-ced2-4308-91fc-e9993667556a-ceph\") pod \"glance-default-internal-api-0\" (UID: \"90a86592-ced2-4308-91fc-e9993667556a\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:54 crc kubenswrapper[4873]: I1201 09:32:54.064309 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4blg\" (UniqueName: \"kubernetes.io/projected/90a86592-ced2-4308-91fc-e9993667556a-kube-api-access-z4blg\") pod \"glance-default-internal-api-0\" (UID: \"90a86592-ced2-4308-91fc-e9993667556a\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:54 crc kubenswrapper[4873]: I1201 09:32:54.064376 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"90a86592-ced2-4308-91fc-e9993667556a\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:54 crc kubenswrapper[4873]: I1201 09:32:54.064449 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/90a86592-ced2-4308-91fc-e9993667556a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"90a86592-ced2-4308-91fc-e9993667556a\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:54 crc kubenswrapper[4873]: I1201 09:32:54.064520 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/90a86592-ced2-4308-91fc-e9993667556a-ceph\") pod \"glance-default-internal-api-0\" (UID: \"90a86592-ced2-4308-91fc-e9993667556a\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:54 crc kubenswrapper[4873]: I1201 09:32:54.064622 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90a86592-ced2-4308-91fc-e9993667556a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"90a86592-ced2-4308-91fc-e9993667556a\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:54 crc kubenswrapper[4873]: I1201 09:32:54.064664 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90a86592-ced2-4308-91fc-e9993667556a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"90a86592-ced2-4308-91fc-e9993667556a\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:54 crc kubenswrapper[4873]: I1201 09:32:54.064681 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90a86592-ced2-4308-91fc-e9993667556a-logs\") pod \"glance-default-internal-api-0\" (UID: \"90a86592-ced2-4308-91fc-e9993667556a\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:54 crc kubenswrapper[4873]: I1201 09:32:54.064746 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90a86592-ced2-4308-91fc-e9993667556a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"90a86592-ced2-4308-91fc-e9993667556a\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:54 crc kubenswrapper[4873]: I1201 09:32:54.064814 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/90a86592-ced2-4308-91fc-e9993667556a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"90a86592-ced2-4308-91fc-e9993667556a\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:54 crc kubenswrapper[4873]: I1201 09:32:54.065425 4873 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"90a86592-ced2-4308-91fc-e9993667556a\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-internal-api-0" Dec 01 09:32:54 crc kubenswrapper[4873]: I1201 09:32:54.065513 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/90a86592-ced2-4308-91fc-e9993667556a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"90a86592-ced2-4308-91fc-e9993667556a\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:54 crc kubenswrapper[4873]: I1201 09:32:54.065802 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90a86592-ced2-4308-91fc-e9993667556a-logs\") pod \"glance-default-internal-api-0\" (UID: \"90a86592-ced2-4308-91fc-e9993667556a\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:54 crc kubenswrapper[4873]: I1201 09:32:54.068862 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/90a86592-ced2-4308-91fc-e9993667556a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"90a86592-ced2-4308-91fc-e9993667556a\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:54 crc kubenswrapper[4873]: I1201 09:32:54.070063 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90a86592-ced2-4308-91fc-e9993667556a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"90a86592-ced2-4308-91fc-e9993667556a\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:54 crc kubenswrapper[4873]: I1201 09:32:54.071161 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90a86592-ced2-4308-91fc-e9993667556a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"90a86592-ced2-4308-91fc-e9993667556a\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:54 crc kubenswrapper[4873]: I1201 09:32:54.071175 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90a86592-ced2-4308-91fc-e9993667556a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"90a86592-ced2-4308-91fc-e9993667556a\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:54 crc kubenswrapper[4873]: I1201 09:32:54.071691 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/90a86592-ced2-4308-91fc-e9993667556a-ceph\") pod \"glance-default-internal-api-0\" (UID: \"90a86592-ced2-4308-91fc-e9993667556a\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:54 crc kubenswrapper[4873]: I1201 09:32:54.088824 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4blg\" (UniqueName: \"kubernetes.io/projected/90a86592-ced2-4308-91fc-e9993667556a-kube-api-access-z4blg\") pod \"glance-default-internal-api-0\" (UID: \"90a86592-ced2-4308-91fc-e9993667556a\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:54 crc kubenswrapper[4873]: I1201 09:32:54.139159 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"90a86592-ced2-4308-91fc-e9993667556a\") " pod="openstack/glance-default-internal-api-0" Dec 01 09:32:54 crc kubenswrapper[4873]: I1201 09:32:54.166853 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 09:32:54 crc kubenswrapper[4873]: I1201 09:32:54.238005 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 09:32:54 crc kubenswrapper[4873]: I1201 09:32:54.451562 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="354e9917-810a-44ba-bc85-b12e15152129" path="/var/lib/kubelet/pods/354e9917-810a-44ba-bc85-b12e15152129/volumes" Dec 01 09:32:54 crc kubenswrapper[4873]: I1201 09:32:54.453251 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5dccf2f5-f149-4b29-99a5-560ff38df6f1" path="/var/lib/kubelet/pods/5dccf2f5-f149-4b29-99a5-560ff38df6f1/volumes" Dec 01 09:32:54 crc kubenswrapper[4873]: I1201 09:32:54.803665 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 09:32:55 crc kubenswrapper[4873]: I1201 09:32:55.176711 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 09:32:55 crc kubenswrapper[4873]: I1201 09:32:55.506236 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"90a86592-ced2-4308-91fc-e9993667556a","Type":"ContainerStarted","Data":"29015f12dcfc291c1505aaeb3659c20b6b9b164c351e0abc9ade551921d37955"} Dec 01 09:32:55 crc kubenswrapper[4873]: I1201 09:32:55.511989 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"380e16b7-2a33-4591-bb99-9382512be92c","Type":"ContainerStarted","Data":"024b85e4a08f5e4c415ff654a7a538a55a3b61bfd7e2aee5b8cf1ea95f5e35a0"} Dec 01 09:32:56 crc kubenswrapper[4873]: I1201 09:32:56.077099 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Dec 01 09:32:56 crc kubenswrapper[4873]: I1201 09:32:56.138819 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-volume1-0" Dec 01 09:32:56 crc kubenswrapper[4873]: I1201 09:32:56.432918 4873 scope.go:117] "RemoveContainer" containerID="1eb6abed9ec75ad80889e13d075ac04aa13db51de45fa2f980efa98223444319" Dec 01 09:32:56 crc kubenswrapper[4873]: E1201 09:32:56.433786 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:32:56 crc kubenswrapper[4873]: I1201 09:32:56.526312 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"90a86592-ced2-4308-91fc-e9993667556a","Type":"ContainerStarted","Data":"9756ecb85b1fa7370e23bcd1ed86e2c8374cfe385b2fc334baa897e8efe9c128"} Dec 01 09:32:57 crc kubenswrapper[4873]: I1201 09:32:57.793878 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-sync-4fq4c"] Dec 01 09:32:57 crc kubenswrapper[4873]: I1201 09:32:57.795870 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-4fq4c" Dec 01 09:32:57 crc kubenswrapper[4873]: I1201 09:32:57.798895 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Dec 01 09:32:57 crc kubenswrapper[4873]: I1201 09:32:57.799527 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-6x5vf" Dec 01 09:32:57 crc kubenswrapper[4873]: I1201 09:32:57.811167 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-4fq4c"] Dec 01 09:32:57 crc kubenswrapper[4873]: I1201 09:32:57.904575 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/f07d0052-de5b-4bc4-9b7b-d65cf86edd06-job-config-data\") pod \"manila-db-sync-4fq4c\" (UID: \"f07d0052-de5b-4bc4-9b7b-d65cf86edd06\") " pod="openstack/manila-db-sync-4fq4c" Dec 01 09:32:57 crc kubenswrapper[4873]: I1201 09:32:57.904889 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f07d0052-de5b-4bc4-9b7b-d65cf86edd06-combined-ca-bundle\") pod \"manila-db-sync-4fq4c\" (UID: \"f07d0052-de5b-4bc4-9b7b-d65cf86edd06\") " pod="openstack/manila-db-sync-4fq4c" Dec 01 09:32:57 crc kubenswrapper[4873]: I1201 09:32:57.905182 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f07d0052-de5b-4bc4-9b7b-d65cf86edd06-config-data\") pod \"manila-db-sync-4fq4c\" (UID: \"f07d0052-de5b-4bc4-9b7b-d65cf86edd06\") " pod="openstack/manila-db-sync-4fq4c" Dec 01 09:32:57 crc kubenswrapper[4873]: I1201 09:32:57.905269 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxb2h\" (UniqueName: \"kubernetes.io/projected/f07d0052-de5b-4bc4-9b7b-d65cf86edd06-kube-api-access-gxb2h\") pod \"manila-db-sync-4fq4c\" (UID: \"f07d0052-de5b-4bc4-9b7b-d65cf86edd06\") " pod="openstack/manila-db-sync-4fq4c" Dec 01 09:32:58 crc kubenswrapper[4873]: I1201 09:32:58.007841 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f07d0052-de5b-4bc4-9b7b-d65cf86edd06-config-data\") pod \"manila-db-sync-4fq4c\" (UID: \"f07d0052-de5b-4bc4-9b7b-d65cf86edd06\") " pod="openstack/manila-db-sync-4fq4c" Dec 01 09:32:58 crc kubenswrapper[4873]: I1201 09:32:58.007927 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxb2h\" (UniqueName: \"kubernetes.io/projected/f07d0052-de5b-4bc4-9b7b-d65cf86edd06-kube-api-access-gxb2h\") pod \"manila-db-sync-4fq4c\" (UID: \"f07d0052-de5b-4bc4-9b7b-d65cf86edd06\") " pod="openstack/manila-db-sync-4fq4c" Dec 01 09:32:58 crc kubenswrapper[4873]: I1201 09:32:58.008034 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/f07d0052-de5b-4bc4-9b7b-d65cf86edd06-job-config-data\") pod \"manila-db-sync-4fq4c\" (UID: \"f07d0052-de5b-4bc4-9b7b-d65cf86edd06\") " pod="openstack/manila-db-sync-4fq4c" Dec 01 09:32:58 crc kubenswrapper[4873]: I1201 09:32:58.008107 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f07d0052-de5b-4bc4-9b7b-d65cf86edd06-combined-ca-bundle\") pod \"manila-db-sync-4fq4c\" (UID: \"f07d0052-de5b-4bc4-9b7b-d65cf86edd06\") " pod="openstack/manila-db-sync-4fq4c" Dec 01 09:32:58 crc kubenswrapper[4873]: I1201 09:32:58.019831 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/f07d0052-de5b-4bc4-9b7b-d65cf86edd06-job-config-data\") pod \"manila-db-sync-4fq4c\" (UID: \"f07d0052-de5b-4bc4-9b7b-d65cf86edd06\") " pod="openstack/manila-db-sync-4fq4c" Dec 01 09:32:58 crc kubenswrapper[4873]: I1201 09:32:58.021159 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f07d0052-de5b-4bc4-9b7b-d65cf86edd06-combined-ca-bundle\") pod \"manila-db-sync-4fq4c\" (UID: \"f07d0052-de5b-4bc4-9b7b-d65cf86edd06\") " pod="openstack/manila-db-sync-4fq4c" Dec 01 09:32:58 crc kubenswrapper[4873]: I1201 09:32:58.026333 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f07d0052-de5b-4bc4-9b7b-d65cf86edd06-config-data\") pod \"manila-db-sync-4fq4c\" (UID: \"f07d0052-de5b-4bc4-9b7b-d65cf86edd06\") " pod="openstack/manila-db-sync-4fq4c" Dec 01 09:32:58 crc kubenswrapper[4873]: I1201 09:32:58.028331 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxb2h\" (UniqueName: \"kubernetes.io/projected/f07d0052-de5b-4bc4-9b7b-d65cf86edd06-kube-api-access-gxb2h\") pod \"manila-db-sync-4fq4c\" (UID: \"f07d0052-de5b-4bc4-9b7b-d65cf86edd06\") " pod="openstack/manila-db-sync-4fq4c" Dec 01 09:32:58 crc kubenswrapper[4873]: I1201 09:32:58.158608 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-4fq4c" Dec 01 09:32:58 crc kubenswrapper[4873]: I1201 09:32:58.550981 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"380e16b7-2a33-4591-bb99-9382512be92c","Type":"ContainerStarted","Data":"b4396dea1dc41f86dc8834f8359b8079eaa57478007a4c99dd6aa123d8e0342d"} Dec 01 09:33:04 crc kubenswrapper[4873]: I1201 09:33:04.143417 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-4fq4c"] Dec 01 09:33:04 crc kubenswrapper[4873]: I1201 09:33:04.630573 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6968fbf848-w68ch" event={"ID":"7f8ecf06-f050-454b-8d36-e91b49847601","Type":"ContainerStarted","Data":"d8579a424e9a6aecde94285d19533b965d7e6a4e298dee09e808fbe0610f8717"} Dec 01 09:33:04 crc kubenswrapper[4873]: I1201 09:33:04.630950 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6968fbf848-w68ch" event={"ID":"7f8ecf06-f050-454b-8d36-e91b49847601","Type":"ContainerStarted","Data":"2c1e990c0c039e99fbb6548784d20ffac2102e30914143bbc13823cbd1f652eb"} Dec 01 09:33:04 crc kubenswrapper[4873]: I1201 09:33:04.632262 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bd6859fdf-xwffw" event={"ID":"a963ef99-5387-4db3-bbe3-b3cedcdb899d","Type":"ContainerStarted","Data":"6ce94b2693845790bf7cfc2937cf542fcb9a8db1291b9be886142a61b88fd38d"} Dec 01 09:33:04 crc kubenswrapper[4873]: I1201 09:33:04.632421 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bd6859fdf-xwffw" event={"ID":"a963ef99-5387-4db3-bbe3-b3cedcdb899d","Type":"ContainerStarted","Data":"453929b758d99b82c99434bbb65d572a5bf6916ea5c7c47d3ebd6c23242f821c"} Dec 01 09:33:04 crc kubenswrapper[4873]: I1201 09:33:04.632538 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7bd6859fdf-xwffw" podUID="a963ef99-5387-4db3-bbe3-b3cedcdb899d" containerName="horizon" containerID="cri-o://6ce94b2693845790bf7cfc2937cf542fcb9a8db1291b9be886142a61b88fd38d" gracePeriod=30 Dec 01 09:33:04 crc kubenswrapper[4873]: I1201 09:33:04.632515 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7bd6859fdf-xwffw" podUID="a963ef99-5387-4db3-bbe3-b3cedcdb899d" containerName="horizon-log" containerID="cri-o://453929b758d99b82c99434bbb65d572a5bf6916ea5c7c47d3ebd6c23242f821c" gracePeriod=30 Dec 01 09:33:04 crc kubenswrapper[4873]: I1201 09:33:04.636775 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"90a86592-ced2-4308-91fc-e9993667556a","Type":"ContainerStarted","Data":"d5f9fcc3f78ae754a341503495ae5f919deb907c40259eedc68dede9d6bdd881"} Dec 01 09:33:04 crc kubenswrapper[4873]: I1201 09:33:04.645894 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c885c5cb8-46fsf" event={"ID":"24452fe2-b990-4537-b619-9db436ed5acb","Type":"ContainerStarted","Data":"19a1f4707bbeeffe7274e84c66714fc06a3e1466289970827846d71126cb57c3"} Dec 01 09:33:04 crc kubenswrapper[4873]: I1201 09:33:04.645956 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c885c5cb8-46fsf" event={"ID":"24452fe2-b990-4537-b619-9db436ed5acb","Type":"ContainerStarted","Data":"3f684c99d8aa2dce24bd09bca0e8fa379cf0d83ddd7621cb5266dddb613a60a8"} Dec 01 09:33:04 crc kubenswrapper[4873]: I1201 09:33:04.650783 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66d489fdc7-jqfsv" event={"ID":"76d98e93-51f4-49f0-93cd-6564618e6304","Type":"ContainerStarted","Data":"f44ef92bf1b8023c2aa0955521d935b25a93df5d0bb925d1afedb91ba70abe9a"} Dec 01 09:33:04 crc kubenswrapper[4873]: I1201 09:33:04.650851 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66d489fdc7-jqfsv" event={"ID":"76d98e93-51f4-49f0-93cd-6564618e6304","Type":"ContainerStarted","Data":"5426af0e98fa5587e33a09bf116d7a4e068bb68eb79d1705830501df2c8917bd"} Dec 01 09:33:04 crc kubenswrapper[4873]: I1201 09:33:04.650899 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-66d489fdc7-jqfsv" podUID="76d98e93-51f4-49f0-93cd-6564618e6304" containerName="horizon-log" containerID="cri-o://5426af0e98fa5587e33a09bf116d7a4e068bb68eb79d1705830501df2c8917bd" gracePeriod=30 Dec 01 09:33:04 crc kubenswrapper[4873]: I1201 09:33:04.650925 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-66d489fdc7-jqfsv" podUID="76d98e93-51f4-49f0-93cd-6564618e6304" containerName="horizon" containerID="cri-o://f44ef92bf1b8023c2aa0955521d935b25a93df5d0bb925d1afedb91ba70abe9a" gracePeriod=30 Dec 01 09:33:04 crc kubenswrapper[4873]: I1201 09:33:04.655566 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-4fq4c" event={"ID":"f07d0052-de5b-4bc4-9b7b-d65cf86edd06","Type":"ContainerStarted","Data":"8543261701bf704a61ed6ef4fb17133afc8c2b834462c0e36432e75322d055ab"} Dec 01 09:33:04 crc kubenswrapper[4873]: I1201 09:33:04.658569 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"380e16b7-2a33-4591-bb99-9382512be92c","Type":"ContainerStarted","Data":"19a65abb2f8ea4f14ca94818b6db55c06cbbcdde9a713ecffdb41e26b3d1de59"} Dec 01 09:33:04 crc kubenswrapper[4873]: I1201 09:33:04.665829 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6968fbf848-w68ch" podStartSLOduration=2.473208378 podStartE2EDuration="14.665800916s" podCreationTimestamp="2025-12-01 09:32:50 +0000 UTC" firstStartedPulling="2025-12-01 09:32:51.453535212 +0000 UTC m=+3147.355643751" lastFinishedPulling="2025-12-01 09:33:03.64612775 +0000 UTC m=+3159.548236289" observedRunningTime="2025-12-01 09:33:04.652448161 +0000 UTC m=+3160.554556710" watchObservedRunningTime="2025-12-01 09:33:04.665800916 +0000 UTC m=+3160.567909455" Dec 01 09:33:04 crc kubenswrapper[4873]: I1201 09:33:04.693942 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-66d489fdc7-jqfsv" podStartSLOduration=2.858852883 podStartE2EDuration="18.69392081s" podCreationTimestamp="2025-12-01 09:32:46 +0000 UTC" firstStartedPulling="2025-12-01 09:32:47.782152328 +0000 UTC m=+3143.684260857" lastFinishedPulling="2025-12-01 09:33:03.617220245 +0000 UTC m=+3159.519328784" observedRunningTime="2025-12-01 09:33:04.69149298 +0000 UTC m=+3160.593601509" watchObservedRunningTime="2025-12-01 09:33:04.69392081 +0000 UTC m=+3160.596029349" Dec 01 09:33:04 crc kubenswrapper[4873]: I1201 09:33:04.720639 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7c885c5cb8-46fsf" podStartSLOduration=3.219992862 podStartE2EDuration="15.720609079s" podCreationTimestamp="2025-12-01 09:32:49 +0000 UTC" firstStartedPulling="2025-12-01 09:32:51.142503487 +0000 UTC m=+3147.044612026" lastFinishedPulling="2025-12-01 09:33:03.643119704 +0000 UTC m=+3159.545228243" observedRunningTime="2025-12-01 09:33:04.710414244 +0000 UTC m=+3160.612522783" watchObservedRunningTime="2025-12-01 09:33:04.720609079 +0000 UTC m=+3160.622717618" Dec 01 09:33:04 crc kubenswrapper[4873]: I1201 09:33:04.752132 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7bd6859fdf-xwffw" podStartSLOduration=2.901455281 podStartE2EDuration="18.752103739s" podCreationTimestamp="2025-12-01 09:32:46 +0000 UTC" firstStartedPulling="2025-12-01 09:32:47.794299342 +0000 UTC m=+3143.696407881" lastFinishedPulling="2025-12-01 09:33:03.6449478 +0000 UTC m=+3159.547056339" observedRunningTime="2025-12-01 09:33:04.740343364 +0000 UTC m=+3160.642451903" watchObservedRunningTime="2025-12-01 09:33:04.752103739 +0000 UTC m=+3160.654212278" Dec 01 09:33:04 crc kubenswrapper[4873]: I1201 09:33:04.773996 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=11.773969727 podStartE2EDuration="11.773969727s" podCreationTimestamp="2025-12-01 09:32:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:33:04.765691449 +0000 UTC m=+3160.667799998" watchObservedRunningTime="2025-12-01 09:33:04.773969727 +0000 UTC m=+3160.676078256" Dec 01 09:33:04 crc kubenswrapper[4873]: I1201 09:33:04.803404 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=11.803380794 podStartE2EDuration="11.803380794s" podCreationTimestamp="2025-12-01 09:32:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:33:04.803056176 +0000 UTC m=+3160.705164715" watchObservedRunningTime="2025-12-01 09:33:04.803380794 +0000 UTC m=+3160.705489333" Dec 01 09:33:06 crc kubenswrapper[4873]: I1201 09:33:06.498958 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7bd6859fdf-xwffw" Dec 01 09:33:07 crc kubenswrapper[4873]: I1201 09:33:07.113204 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-66d489fdc7-jqfsv" Dec 01 09:33:08 crc kubenswrapper[4873]: I1201 09:33:08.430300 4873 scope.go:117] "RemoveContainer" containerID="1eb6abed9ec75ad80889e13d075ac04aa13db51de45fa2f980efa98223444319" Dec 01 09:33:08 crc kubenswrapper[4873]: E1201 09:33:08.431004 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:33:10 crc kubenswrapper[4873]: I1201 09:33:10.346831 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7c885c5cb8-46fsf" Dec 01 09:33:10 crc kubenswrapper[4873]: I1201 09:33:10.347283 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7c885c5cb8-46fsf" Dec 01 09:33:10 crc kubenswrapper[4873]: I1201 09:33:10.660646 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6968fbf848-w68ch" Dec 01 09:33:10 crc kubenswrapper[4873]: I1201 09:33:10.660715 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6968fbf848-w68ch" Dec 01 09:33:13 crc kubenswrapper[4873]: I1201 09:33:13.781368 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-4fq4c" event={"ID":"f07d0052-de5b-4bc4-9b7b-d65cf86edd06","Type":"ContainerStarted","Data":"839143e7968b30f668bf8c94fe6e5226bd3dc152c8426bde3713d5f48d698806"} Dec 01 09:33:14 crc kubenswrapper[4873]: I1201 09:33:14.167796 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 01 09:33:14 crc kubenswrapper[4873]: I1201 09:33:14.168130 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 01 09:33:14 crc kubenswrapper[4873]: I1201 09:33:14.216483 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 01 09:33:14 crc kubenswrapper[4873]: I1201 09:33:14.238500 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 01 09:33:14 crc kubenswrapper[4873]: I1201 09:33:14.238553 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 01 09:33:14 crc kubenswrapper[4873]: I1201 09:33:14.243848 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 01 09:33:14 crc kubenswrapper[4873]: I1201 09:33:14.252542 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-sync-4fq4c" podStartSLOduration=8.238956431 podStartE2EDuration="17.252517084s" podCreationTimestamp="2025-12-01 09:32:57 +0000 UTC" firstStartedPulling="2025-12-01 09:33:04.183905538 +0000 UTC m=+3160.086014077" lastFinishedPulling="2025-12-01 09:33:13.197466191 +0000 UTC m=+3169.099574730" observedRunningTime="2025-12-01 09:33:13.803643654 +0000 UTC m=+3169.705752223" watchObservedRunningTime="2025-12-01 09:33:14.252517084 +0000 UTC m=+3170.154625623" Dec 01 09:33:14 crc kubenswrapper[4873]: I1201 09:33:14.279611 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 01 09:33:14 crc kubenswrapper[4873]: I1201 09:33:14.312166 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 01 09:33:14 crc kubenswrapper[4873]: I1201 09:33:14.793293 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 01 09:33:14 crc kubenswrapper[4873]: I1201 09:33:14.793529 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 01 09:33:14 crc kubenswrapper[4873]: I1201 09:33:14.793779 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 01 09:33:14 crc kubenswrapper[4873]: I1201 09:33:14.793800 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 01 09:33:17 crc kubenswrapper[4873]: I1201 09:33:17.638775 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 01 09:33:17 crc kubenswrapper[4873]: I1201 09:33:17.640584 4873 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 09:33:17 crc kubenswrapper[4873]: I1201 09:33:17.646332 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 01 09:33:17 crc kubenswrapper[4873]: I1201 09:33:17.884157 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 01 09:33:17 crc kubenswrapper[4873]: I1201 09:33:17.884303 4873 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 09:33:17 crc kubenswrapper[4873]: I1201 09:33:17.888612 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 01 09:33:20 crc kubenswrapper[4873]: I1201 09:33:20.350371 4873 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7c885c5cb8-46fsf" podUID="24452fe2-b990-4537-b619-9db436ed5acb" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.242:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.242:8443: connect: connection refused" Dec 01 09:33:20 crc kubenswrapper[4873]: I1201 09:33:20.663903 4873 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6968fbf848-w68ch" podUID="7f8ecf06-f050-454b-8d36-e91b49847601" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.243:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.243:8443: connect: connection refused" Dec 01 09:33:23 crc kubenswrapper[4873]: I1201 09:33:23.430833 4873 scope.go:117] "RemoveContainer" containerID="1eb6abed9ec75ad80889e13d075ac04aa13db51de45fa2f980efa98223444319" Dec 01 09:33:23 crc kubenswrapper[4873]: E1201 09:33:23.432222 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:33:28 crc kubenswrapper[4873]: I1201 09:33:28.004456 4873 generic.go:334] "Generic (PLEG): container finished" podID="f07d0052-de5b-4bc4-9b7b-d65cf86edd06" containerID="839143e7968b30f668bf8c94fe6e5226bd3dc152c8426bde3713d5f48d698806" exitCode=0 Dec 01 09:33:28 crc kubenswrapper[4873]: I1201 09:33:28.004538 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-4fq4c" event={"ID":"f07d0052-de5b-4bc4-9b7b-d65cf86edd06","Type":"ContainerDied","Data":"839143e7968b30f668bf8c94fe6e5226bd3dc152c8426bde3713d5f48d698806"} Dec 01 09:33:29 crc kubenswrapper[4873]: I1201 09:33:29.541250 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-4fq4c" Dec 01 09:33:29 crc kubenswrapper[4873]: I1201 09:33:29.639192 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f07d0052-de5b-4bc4-9b7b-d65cf86edd06-config-data\") pod \"f07d0052-de5b-4bc4-9b7b-d65cf86edd06\" (UID: \"f07d0052-de5b-4bc4-9b7b-d65cf86edd06\") " Dec 01 09:33:29 crc kubenswrapper[4873]: I1201 09:33:29.639360 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/f07d0052-de5b-4bc4-9b7b-d65cf86edd06-job-config-data\") pod \"f07d0052-de5b-4bc4-9b7b-d65cf86edd06\" (UID: \"f07d0052-de5b-4bc4-9b7b-d65cf86edd06\") " Dec 01 09:33:29 crc kubenswrapper[4873]: I1201 09:33:29.639557 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f07d0052-de5b-4bc4-9b7b-d65cf86edd06-combined-ca-bundle\") pod \"f07d0052-de5b-4bc4-9b7b-d65cf86edd06\" (UID: \"f07d0052-de5b-4bc4-9b7b-d65cf86edd06\") " Dec 01 09:33:29 crc kubenswrapper[4873]: I1201 09:33:29.639827 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gxb2h\" (UniqueName: \"kubernetes.io/projected/f07d0052-de5b-4bc4-9b7b-d65cf86edd06-kube-api-access-gxb2h\") pod \"f07d0052-de5b-4bc4-9b7b-d65cf86edd06\" (UID: \"f07d0052-de5b-4bc4-9b7b-d65cf86edd06\") " Dec 01 09:33:29 crc kubenswrapper[4873]: I1201 09:33:29.647401 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f07d0052-de5b-4bc4-9b7b-d65cf86edd06-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "f07d0052-de5b-4bc4-9b7b-d65cf86edd06" (UID: "f07d0052-de5b-4bc4-9b7b-d65cf86edd06"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:33:29 crc kubenswrapper[4873]: I1201 09:33:29.651312 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f07d0052-de5b-4bc4-9b7b-d65cf86edd06-config-data" (OuterVolumeSpecName: "config-data") pod "f07d0052-de5b-4bc4-9b7b-d65cf86edd06" (UID: "f07d0052-de5b-4bc4-9b7b-d65cf86edd06"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:33:29 crc kubenswrapper[4873]: I1201 09:33:29.663011 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f07d0052-de5b-4bc4-9b7b-d65cf86edd06-kube-api-access-gxb2h" (OuterVolumeSpecName: "kube-api-access-gxb2h") pod "f07d0052-de5b-4bc4-9b7b-d65cf86edd06" (UID: "f07d0052-de5b-4bc4-9b7b-d65cf86edd06"). InnerVolumeSpecName "kube-api-access-gxb2h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:33:29 crc kubenswrapper[4873]: I1201 09:33:29.678680 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f07d0052-de5b-4bc4-9b7b-d65cf86edd06-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f07d0052-de5b-4bc4-9b7b-d65cf86edd06" (UID: "f07d0052-de5b-4bc4-9b7b-d65cf86edd06"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:33:29 crc kubenswrapper[4873]: I1201 09:33:29.743262 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gxb2h\" (UniqueName: \"kubernetes.io/projected/f07d0052-de5b-4bc4-9b7b-d65cf86edd06-kube-api-access-gxb2h\") on node \"crc\" DevicePath \"\"" Dec 01 09:33:29 crc kubenswrapper[4873]: I1201 09:33:29.743312 4873 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f07d0052-de5b-4bc4-9b7b-d65cf86edd06-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:33:29 crc kubenswrapper[4873]: I1201 09:33:29.743328 4873 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/f07d0052-de5b-4bc4-9b7b-d65cf86edd06-job-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:33:29 crc kubenswrapper[4873]: I1201 09:33:29.743338 4873 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f07d0052-de5b-4bc4-9b7b-d65cf86edd06-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.050273 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-4fq4c" event={"ID":"f07d0052-de5b-4bc4-9b7b-d65cf86edd06","Type":"ContainerDied","Data":"8543261701bf704a61ed6ef4fb17133afc8c2b834462c0e36432e75322d055ab"} Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.050327 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8543261701bf704a61ed6ef4fb17133afc8c2b834462c0e36432e75322d055ab" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.050353 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-4fq4c" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.476174 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Dec 01 09:33:30 crc kubenswrapper[4873]: E1201 09:33:30.477073 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f07d0052-de5b-4bc4-9b7b-d65cf86edd06" containerName="manila-db-sync" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.477092 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="f07d0052-de5b-4bc4-9b7b-d65cf86edd06" containerName="manila-db-sync" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.477344 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="f07d0052-de5b-4bc4-9b7b-d65cf86edd06" containerName="manila-db-sync" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.479094 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.514610 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.514831 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.515086 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-6x5vf" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.515295 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scripts" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.535140 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.537592 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.544344 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.546408 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.562850 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67d40a29-6c05-4dfa-9caa-457176a83b49-config-data\") pod \"manila-share-share1-0\" (UID: \"67d40a29-6c05-4dfa-9caa-457176a83b49\") " pod="openstack/manila-share-share1-0" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.562905 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15f823d3-d00f-4c94-bd11-3cd81c911b10-config-data\") pod \"manila-scheduler-0\" (UID: \"15f823d3-d00f-4c94-bd11-3cd81c911b10\") " pod="openstack/manila-scheduler-0" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.562928 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rv9gc\" (UniqueName: \"kubernetes.io/projected/15f823d3-d00f-4c94-bd11-3cd81c911b10-kube-api-access-rv9gc\") pod \"manila-scheduler-0\" (UID: \"15f823d3-d00f-4c94-bd11-3cd81c911b10\") " pod="openstack/manila-scheduler-0" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.562947 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/67d40a29-6c05-4dfa-9caa-457176a83b49-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"67d40a29-6c05-4dfa-9caa-457176a83b49\") " pod="openstack/manila-share-share1-0" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.563055 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15f823d3-d00f-4c94-bd11-3cd81c911b10-scripts\") pod \"manila-scheduler-0\" (UID: \"15f823d3-d00f-4c94-bd11-3cd81c911b10\") " pod="openstack/manila-scheduler-0" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.563109 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/67d40a29-6c05-4dfa-9caa-457176a83b49-ceph\") pod \"manila-share-share1-0\" (UID: \"67d40a29-6c05-4dfa-9caa-457176a83b49\") " pod="openstack/manila-share-share1-0" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.563128 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/15f823d3-d00f-4c94-bd11-3cd81c911b10-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"15f823d3-d00f-4c94-bd11-3cd81c911b10\") " pod="openstack/manila-scheduler-0" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.563161 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/15f823d3-d00f-4c94-bd11-3cd81c911b10-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"15f823d3-d00f-4c94-bd11-3cd81c911b10\") " pod="openstack/manila-scheduler-0" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.563258 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67d40a29-6c05-4dfa-9caa-457176a83b49-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"67d40a29-6c05-4dfa-9caa-457176a83b49\") " pod="openstack/manila-share-share1-0" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.563284 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/67d40a29-6c05-4dfa-9caa-457176a83b49-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"67d40a29-6c05-4dfa-9caa-457176a83b49\") " pod="openstack/manila-share-share1-0" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.563304 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15f823d3-d00f-4c94-bd11-3cd81c911b10-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"15f823d3-d00f-4c94-bd11-3cd81c911b10\") " pod="openstack/manila-scheduler-0" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.563368 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67d40a29-6c05-4dfa-9caa-457176a83b49-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"67d40a29-6c05-4dfa-9caa-457176a83b49\") " pod="openstack/manila-share-share1-0" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.563391 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67d40a29-6c05-4dfa-9caa-457176a83b49-scripts\") pod \"manila-share-share1-0\" (UID: \"67d40a29-6c05-4dfa-9caa-457176a83b49\") " pod="openstack/manila-share-share1-0" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.563410 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8h49t\" (UniqueName: \"kubernetes.io/projected/67d40a29-6c05-4dfa-9caa-457176a83b49-kube-api-access-8h49t\") pod \"manila-share-share1-0\" (UID: \"67d40a29-6c05-4dfa-9caa-457176a83b49\") " pod="openstack/manila-share-share1-0" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.604991 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.665858 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67d40a29-6c05-4dfa-9caa-457176a83b49-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"67d40a29-6c05-4dfa-9caa-457176a83b49\") " pod="openstack/manila-share-share1-0" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.665922 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/67d40a29-6c05-4dfa-9caa-457176a83b49-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"67d40a29-6c05-4dfa-9caa-457176a83b49\") " pod="openstack/manila-share-share1-0" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.665943 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15f823d3-d00f-4c94-bd11-3cd81c911b10-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"15f823d3-d00f-4c94-bd11-3cd81c911b10\") " pod="openstack/manila-scheduler-0" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.665994 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67d40a29-6c05-4dfa-9caa-457176a83b49-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"67d40a29-6c05-4dfa-9caa-457176a83b49\") " pod="openstack/manila-share-share1-0" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.666028 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67d40a29-6c05-4dfa-9caa-457176a83b49-scripts\") pod \"manila-share-share1-0\" (UID: \"67d40a29-6c05-4dfa-9caa-457176a83b49\") " pod="openstack/manila-share-share1-0" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.666054 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8h49t\" (UniqueName: \"kubernetes.io/projected/67d40a29-6c05-4dfa-9caa-457176a83b49-kube-api-access-8h49t\") pod \"manila-share-share1-0\" (UID: \"67d40a29-6c05-4dfa-9caa-457176a83b49\") " pod="openstack/manila-share-share1-0" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.666097 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67d40a29-6c05-4dfa-9caa-457176a83b49-config-data\") pod \"manila-share-share1-0\" (UID: \"67d40a29-6c05-4dfa-9caa-457176a83b49\") " pod="openstack/manila-share-share1-0" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.666117 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15f823d3-d00f-4c94-bd11-3cd81c911b10-config-data\") pod \"manila-scheduler-0\" (UID: \"15f823d3-d00f-4c94-bd11-3cd81c911b10\") " pod="openstack/manila-scheduler-0" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.666139 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rv9gc\" (UniqueName: \"kubernetes.io/projected/15f823d3-d00f-4c94-bd11-3cd81c911b10-kube-api-access-rv9gc\") pod \"manila-scheduler-0\" (UID: \"15f823d3-d00f-4c94-bd11-3cd81c911b10\") " pod="openstack/manila-scheduler-0" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.666161 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/67d40a29-6c05-4dfa-9caa-457176a83b49-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"67d40a29-6c05-4dfa-9caa-457176a83b49\") " pod="openstack/manila-share-share1-0" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.666225 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15f823d3-d00f-4c94-bd11-3cd81c911b10-scripts\") pod \"manila-scheduler-0\" (UID: \"15f823d3-d00f-4c94-bd11-3cd81c911b10\") " pod="openstack/manila-scheduler-0" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.666272 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/67d40a29-6c05-4dfa-9caa-457176a83b49-ceph\") pod \"manila-share-share1-0\" (UID: \"67d40a29-6c05-4dfa-9caa-457176a83b49\") " pod="openstack/manila-share-share1-0" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.666296 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/15f823d3-d00f-4c94-bd11-3cd81c911b10-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"15f823d3-d00f-4c94-bd11-3cd81c911b10\") " pod="openstack/manila-scheduler-0" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.666325 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/15f823d3-d00f-4c94-bd11-3cd81c911b10-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"15f823d3-d00f-4c94-bd11-3cd81c911b10\") " pod="openstack/manila-scheduler-0" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.666470 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/15f823d3-d00f-4c94-bd11-3cd81c911b10-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"15f823d3-d00f-4c94-bd11-3cd81c911b10\") " pod="openstack/manila-scheduler-0" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.668494 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/67d40a29-6c05-4dfa-9caa-457176a83b49-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"67d40a29-6c05-4dfa-9caa-457176a83b49\") " pod="openstack/manila-share-share1-0" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.672950 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67d40a29-6c05-4dfa-9caa-457176a83b49-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"67d40a29-6c05-4dfa-9caa-457176a83b49\") " pod="openstack/manila-share-share1-0" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.674156 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/67d40a29-6c05-4dfa-9caa-457176a83b49-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"67d40a29-6c05-4dfa-9caa-457176a83b49\") " pod="openstack/manila-share-share1-0" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.687941 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67d40a29-6c05-4dfa-9caa-457176a83b49-scripts\") pod \"manila-share-share1-0\" (UID: \"67d40a29-6c05-4dfa-9caa-457176a83b49\") " pod="openstack/manila-share-share1-0" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.688814 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67d40a29-6c05-4dfa-9caa-457176a83b49-config-data\") pod \"manila-share-share1-0\" (UID: \"67d40a29-6c05-4dfa-9caa-457176a83b49\") " pod="openstack/manila-share-share1-0" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.692767 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15f823d3-d00f-4c94-bd11-3cd81c911b10-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"15f823d3-d00f-4c94-bd11-3cd81c911b10\") " pod="openstack/manila-scheduler-0" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.693320 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15f823d3-d00f-4c94-bd11-3cd81c911b10-config-data\") pod \"manila-scheduler-0\" (UID: \"15f823d3-d00f-4c94-bd11-3cd81c911b10\") " pod="openstack/manila-scheduler-0" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.695167 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67d40a29-6c05-4dfa-9caa-457176a83b49-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"67d40a29-6c05-4dfa-9caa-457176a83b49\") " pod="openstack/manila-share-share1-0" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.695800 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/15f823d3-d00f-4c94-bd11-3cd81c911b10-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"15f823d3-d00f-4c94-bd11-3cd81c911b10\") " pod="openstack/manila-scheduler-0" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.699169 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/67d40a29-6c05-4dfa-9caa-457176a83b49-ceph\") pod \"manila-share-share1-0\" (UID: \"67d40a29-6c05-4dfa-9caa-457176a83b49\") " pod="openstack/manila-share-share1-0" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.714427 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15f823d3-d00f-4c94-bd11-3cd81c911b10-scripts\") pod \"manila-scheduler-0\" (UID: \"15f823d3-d00f-4c94-bd11-3cd81c911b10\") " pod="openstack/manila-scheduler-0" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.724432 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rv9gc\" (UniqueName: \"kubernetes.io/projected/15f823d3-d00f-4c94-bd11-3cd81c911b10-kube-api-access-rv9gc\") pod \"manila-scheduler-0\" (UID: \"15f823d3-d00f-4c94-bd11-3cd81c911b10\") " pod="openstack/manila-scheduler-0" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.733517 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8h49t\" (UniqueName: \"kubernetes.io/projected/67d40a29-6c05-4dfa-9caa-457176a83b49-kube-api-access-8h49t\") pod \"manila-share-share1-0\" (UID: \"67d40a29-6c05-4dfa-9caa-457176a83b49\") " pod="openstack/manila-share-share1-0" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.831671 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.864413 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.894432 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-76b5fdb995-hdbhs"] Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.904293 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76b5fdb995-hdbhs" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.951484 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76b5fdb995-hdbhs"] Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.990148 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/90639568-2248-4249-be34-140ed23a5d1d-openstack-edpm-ipam\") pod \"dnsmasq-dns-76b5fdb995-hdbhs\" (UID: \"90639568-2248-4249-be34-140ed23a5d1d\") " pod="openstack/dnsmasq-dns-76b5fdb995-hdbhs" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.990246 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90639568-2248-4249-be34-140ed23a5d1d-config\") pod \"dnsmasq-dns-76b5fdb995-hdbhs\" (UID: \"90639568-2248-4249-be34-140ed23a5d1d\") " pod="openstack/dnsmasq-dns-76b5fdb995-hdbhs" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.990281 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/90639568-2248-4249-be34-140ed23a5d1d-ovsdbserver-nb\") pod \"dnsmasq-dns-76b5fdb995-hdbhs\" (UID: \"90639568-2248-4249-be34-140ed23a5d1d\") " pod="openstack/dnsmasq-dns-76b5fdb995-hdbhs" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.990327 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljrzb\" (UniqueName: \"kubernetes.io/projected/90639568-2248-4249-be34-140ed23a5d1d-kube-api-access-ljrzb\") pod \"dnsmasq-dns-76b5fdb995-hdbhs\" (UID: \"90639568-2248-4249-be34-140ed23a5d1d\") " pod="openstack/dnsmasq-dns-76b5fdb995-hdbhs" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.990375 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/90639568-2248-4249-be34-140ed23a5d1d-dns-svc\") pod \"dnsmasq-dns-76b5fdb995-hdbhs\" (UID: \"90639568-2248-4249-be34-140ed23a5d1d\") " pod="openstack/dnsmasq-dns-76b5fdb995-hdbhs" Dec 01 09:33:30 crc kubenswrapper[4873]: I1201 09:33:30.990477 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/90639568-2248-4249-be34-140ed23a5d1d-ovsdbserver-sb\") pod \"dnsmasq-dns-76b5fdb995-hdbhs\" (UID: \"90639568-2248-4249-be34-140ed23a5d1d\") " pod="openstack/dnsmasq-dns-76b5fdb995-hdbhs" Dec 01 09:33:31 crc kubenswrapper[4873]: I1201 09:33:31.100341 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/90639568-2248-4249-be34-140ed23a5d1d-openstack-edpm-ipam\") pod \"dnsmasq-dns-76b5fdb995-hdbhs\" (UID: \"90639568-2248-4249-be34-140ed23a5d1d\") " pod="openstack/dnsmasq-dns-76b5fdb995-hdbhs" Dec 01 09:33:31 crc kubenswrapper[4873]: I1201 09:33:31.101136 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90639568-2248-4249-be34-140ed23a5d1d-config\") pod \"dnsmasq-dns-76b5fdb995-hdbhs\" (UID: \"90639568-2248-4249-be34-140ed23a5d1d\") " pod="openstack/dnsmasq-dns-76b5fdb995-hdbhs" Dec 01 09:33:31 crc kubenswrapper[4873]: I1201 09:33:31.101616 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/90639568-2248-4249-be34-140ed23a5d1d-ovsdbserver-nb\") pod \"dnsmasq-dns-76b5fdb995-hdbhs\" (UID: \"90639568-2248-4249-be34-140ed23a5d1d\") " pod="openstack/dnsmasq-dns-76b5fdb995-hdbhs" Dec 01 09:33:31 crc kubenswrapper[4873]: I1201 09:33:31.101514 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/90639568-2248-4249-be34-140ed23a5d1d-openstack-edpm-ipam\") pod \"dnsmasq-dns-76b5fdb995-hdbhs\" (UID: \"90639568-2248-4249-be34-140ed23a5d1d\") " pod="openstack/dnsmasq-dns-76b5fdb995-hdbhs" Dec 01 09:33:31 crc kubenswrapper[4873]: I1201 09:33:31.101745 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljrzb\" (UniqueName: \"kubernetes.io/projected/90639568-2248-4249-be34-140ed23a5d1d-kube-api-access-ljrzb\") pod \"dnsmasq-dns-76b5fdb995-hdbhs\" (UID: \"90639568-2248-4249-be34-140ed23a5d1d\") " pod="openstack/dnsmasq-dns-76b5fdb995-hdbhs" Dec 01 09:33:31 crc kubenswrapper[4873]: I1201 09:33:31.102662 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/90639568-2248-4249-be34-140ed23a5d1d-ovsdbserver-nb\") pod \"dnsmasq-dns-76b5fdb995-hdbhs\" (UID: \"90639568-2248-4249-be34-140ed23a5d1d\") " pod="openstack/dnsmasq-dns-76b5fdb995-hdbhs" Dec 01 09:33:31 crc kubenswrapper[4873]: I1201 09:33:31.103128 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/90639568-2248-4249-be34-140ed23a5d1d-dns-svc\") pod \"dnsmasq-dns-76b5fdb995-hdbhs\" (UID: \"90639568-2248-4249-be34-140ed23a5d1d\") " pod="openstack/dnsmasq-dns-76b5fdb995-hdbhs" Dec 01 09:33:31 crc kubenswrapper[4873]: I1201 09:33:31.103381 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/90639568-2248-4249-be34-140ed23a5d1d-ovsdbserver-sb\") pod \"dnsmasq-dns-76b5fdb995-hdbhs\" (UID: \"90639568-2248-4249-be34-140ed23a5d1d\") " pod="openstack/dnsmasq-dns-76b5fdb995-hdbhs" Dec 01 09:33:31 crc kubenswrapper[4873]: I1201 09:33:31.104480 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/90639568-2248-4249-be34-140ed23a5d1d-dns-svc\") pod \"dnsmasq-dns-76b5fdb995-hdbhs\" (UID: \"90639568-2248-4249-be34-140ed23a5d1d\") " pod="openstack/dnsmasq-dns-76b5fdb995-hdbhs" Dec 01 09:33:31 crc kubenswrapper[4873]: I1201 09:33:31.105208 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/90639568-2248-4249-be34-140ed23a5d1d-ovsdbserver-sb\") pod \"dnsmasq-dns-76b5fdb995-hdbhs\" (UID: \"90639568-2248-4249-be34-140ed23a5d1d\") " pod="openstack/dnsmasq-dns-76b5fdb995-hdbhs" Dec 01 09:33:31 crc kubenswrapper[4873]: I1201 09:33:31.103383 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Dec 01 09:33:31 crc kubenswrapper[4873]: I1201 09:33:31.110537 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 01 09:33:31 crc kubenswrapper[4873]: I1201 09:33:31.122552 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Dec 01 09:33:31 crc kubenswrapper[4873]: I1201 09:33:31.130793 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90639568-2248-4249-be34-140ed23a5d1d-config\") pod \"dnsmasq-dns-76b5fdb995-hdbhs\" (UID: \"90639568-2248-4249-be34-140ed23a5d1d\") " pod="openstack/dnsmasq-dns-76b5fdb995-hdbhs" Dec 01 09:33:31 crc kubenswrapper[4873]: I1201 09:33:31.167033 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Dec 01 09:33:31 crc kubenswrapper[4873]: I1201 09:33:31.202104 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljrzb\" (UniqueName: \"kubernetes.io/projected/90639568-2248-4249-be34-140ed23a5d1d-kube-api-access-ljrzb\") pod \"dnsmasq-dns-76b5fdb995-hdbhs\" (UID: \"90639568-2248-4249-be34-140ed23a5d1d\") " pod="openstack/dnsmasq-dns-76b5fdb995-hdbhs" Dec 01 09:33:31 crc kubenswrapper[4873]: I1201 09:33:31.231848 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8458dfc6-db77-4d48-9c9a-9324813002a3-config-data-custom\") pod \"manila-api-0\" (UID: \"8458dfc6-db77-4d48-9c9a-9324813002a3\") " pod="openstack/manila-api-0" Dec 01 09:33:31 crc kubenswrapper[4873]: I1201 09:33:31.231981 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6zh7\" (UniqueName: \"kubernetes.io/projected/8458dfc6-db77-4d48-9c9a-9324813002a3-kube-api-access-s6zh7\") pod \"manila-api-0\" (UID: \"8458dfc6-db77-4d48-9c9a-9324813002a3\") " pod="openstack/manila-api-0" Dec 01 09:33:31 crc kubenswrapper[4873]: I1201 09:33:31.232003 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8458dfc6-db77-4d48-9c9a-9324813002a3-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"8458dfc6-db77-4d48-9c9a-9324813002a3\") " pod="openstack/manila-api-0" Dec 01 09:33:31 crc kubenswrapper[4873]: I1201 09:33:31.232065 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8458dfc6-db77-4d48-9c9a-9324813002a3-scripts\") pod \"manila-api-0\" (UID: \"8458dfc6-db77-4d48-9c9a-9324813002a3\") " pod="openstack/manila-api-0" Dec 01 09:33:31 crc kubenswrapper[4873]: I1201 09:33:31.232102 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8458dfc6-db77-4d48-9c9a-9324813002a3-config-data\") pod \"manila-api-0\" (UID: \"8458dfc6-db77-4d48-9c9a-9324813002a3\") " pod="openstack/manila-api-0" Dec 01 09:33:31 crc kubenswrapper[4873]: I1201 09:33:31.232140 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8458dfc6-db77-4d48-9c9a-9324813002a3-etc-machine-id\") pod \"manila-api-0\" (UID: \"8458dfc6-db77-4d48-9c9a-9324813002a3\") " pod="openstack/manila-api-0" Dec 01 09:33:31 crc kubenswrapper[4873]: I1201 09:33:31.232159 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8458dfc6-db77-4d48-9c9a-9324813002a3-logs\") pod \"manila-api-0\" (UID: \"8458dfc6-db77-4d48-9c9a-9324813002a3\") " pod="openstack/manila-api-0" Dec 01 09:33:31 crc kubenswrapper[4873]: I1201 09:33:31.370393 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8458dfc6-db77-4d48-9c9a-9324813002a3-config-data-custom\") pod \"manila-api-0\" (UID: \"8458dfc6-db77-4d48-9c9a-9324813002a3\") " pod="openstack/manila-api-0" Dec 01 09:33:31 crc kubenswrapper[4873]: I1201 09:33:31.370607 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6zh7\" (UniqueName: \"kubernetes.io/projected/8458dfc6-db77-4d48-9c9a-9324813002a3-kube-api-access-s6zh7\") pod \"manila-api-0\" (UID: \"8458dfc6-db77-4d48-9c9a-9324813002a3\") " pod="openstack/manila-api-0" Dec 01 09:33:31 crc kubenswrapper[4873]: I1201 09:33:31.370645 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8458dfc6-db77-4d48-9c9a-9324813002a3-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"8458dfc6-db77-4d48-9c9a-9324813002a3\") " pod="openstack/manila-api-0" Dec 01 09:33:31 crc kubenswrapper[4873]: I1201 09:33:31.370720 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8458dfc6-db77-4d48-9c9a-9324813002a3-scripts\") pod \"manila-api-0\" (UID: \"8458dfc6-db77-4d48-9c9a-9324813002a3\") " pod="openstack/manila-api-0" Dec 01 09:33:31 crc kubenswrapper[4873]: I1201 09:33:31.370784 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8458dfc6-db77-4d48-9c9a-9324813002a3-config-data\") pod \"manila-api-0\" (UID: \"8458dfc6-db77-4d48-9c9a-9324813002a3\") " pod="openstack/manila-api-0" Dec 01 09:33:31 crc kubenswrapper[4873]: I1201 09:33:31.370846 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8458dfc6-db77-4d48-9c9a-9324813002a3-etc-machine-id\") pod \"manila-api-0\" (UID: \"8458dfc6-db77-4d48-9c9a-9324813002a3\") " pod="openstack/manila-api-0" Dec 01 09:33:31 crc kubenswrapper[4873]: I1201 09:33:31.370877 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8458dfc6-db77-4d48-9c9a-9324813002a3-logs\") pod \"manila-api-0\" (UID: \"8458dfc6-db77-4d48-9c9a-9324813002a3\") " pod="openstack/manila-api-0" Dec 01 09:33:31 crc kubenswrapper[4873]: I1201 09:33:31.372037 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8458dfc6-db77-4d48-9c9a-9324813002a3-logs\") pod \"manila-api-0\" (UID: \"8458dfc6-db77-4d48-9c9a-9324813002a3\") " pod="openstack/manila-api-0" Dec 01 09:33:31 crc kubenswrapper[4873]: I1201 09:33:31.373326 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8458dfc6-db77-4d48-9c9a-9324813002a3-etc-machine-id\") pod \"manila-api-0\" (UID: \"8458dfc6-db77-4d48-9c9a-9324813002a3\") " pod="openstack/manila-api-0" Dec 01 09:33:31 crc kubenswrapper[4873]: I1201 09:33:31.384855 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8458dfc6-db77-4d48-9c9a-9324813002a3-scripts\") pod \"manila-api-0\" (UID: \"8458dfc6-db77-4d48-9c9a-9324813002a3\") " pod="openstack/manila-api-0" Dec 01 09:33:31 crc kubenswrapper[4873]: I1201 09:33:31.387516 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76b5fdb995-hdbhs" Dec 01 09:33:31 crc kubenswrapper[4873]: I1201 09:33:31.389981 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8458dfc6-db77-4d48-9c9a-9324813002a3-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"8458dfc6-db77-4d48-9c9a-9324813002a3\") " pod="openstack/manila-api-0" Dec 01 09:33:31 crc kubenswrapper[4873]: I1201 09:33:31.391662 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8458dfc6-db77-4d48-9c9a-9324813002a3-config-data-custom\") pod \"manila-api-0\" (UID: \"8458dfc6-db77-4d48-9c9a-9324813002a3\") " pod="openstack/manila-api-0" Dec 01 09:33:31 crc kubenswrapper[4873]: I1201 09:33:31.407855 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8458dfc6-db77-4d48-9c9a-9324813002a3-config-data\") pod \"manila-api-0\" (UID: \"8458dfc6-db77-4d48-9c9a-9324813002a3\") " pod="openstack/manila-api-0" Dec 01 09:33:31 crc kubenswrapper[4873]: I1201 09:33:31.425195 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6zh7\" (UniqueName: \"kubernetes.io/projected/8458dfc6-db77-4d48-9c9a-9324813002a3-kube-api-access-s6zh7\") pod \"manila-api-0\" (UID: \"8458dfc6-db77-4d48-9c9a-9324813002a3\") " pod="openstack/manila-api-0" Dec 01 09:33:31 crc kubenswrapper[4873]: I1201 09:33:31.461093 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 01 09:33:31 crc kubenswrapper[4873]: I1201 09:33:31.844721 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Dec 01 09:33:32 crc kubenswrapper[4873]: I1201 09:33:32.116343 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"15f823d3-d00f-4c94-bd11-3cd81c911b10","Type":"ContainerStarted","Data":"403f125ce7d0ad764382436581a0b6386c4ef4a5b9d96cc03720e541176939c5"} Dec 01 09:33:32 crc kubenswrapper[4873]: I1201 09:33:32.226069 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Dec 01 09:33:32 crc kubenswrapper[4873]: I1201 09:33:32.250750 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76b5fdb995-hdbhs"] Dec 01 09:33:32 crc kubenswrapper[4873]: I1201 09:33:32.403360 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Dec 01 09:33:33 crc kubenswrapper[4873]: I1201 09:33:33.151108 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"67d40a29-6c05-4dfa-9caa-457176a83b49","Type":"ContainerStarted","Data":"3c37578b4a689e47bba8e7ab299ba0e86eb630ae21dca7997fa71c0547581d8a"} Dec 01 09:33:33 crc kubenswrapper[4873]: I1201 09:33:33.155794 4873 generic.go:334] "Generic (PLEG): container finished" podID="90639568-2248-4249-be34-140ed23a5d1d" containerID="cd2c555f751777465746f116a7bca162eeecc5592f25d4143a45b860290d30e2" exitCode=0 Dec 01 09:33:33 crc kubenswrapper[4873]: I1201 09:33:33.155852 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76b5fdb995-hdbhs" event={"ID":"90639568-2248-4249-be34-140ed23a5d1d","Type":"ContainerDied","Data":"cd2c555f751777465746f116a7bca162eeecc5592f25d4143a45b860290d30e2"} Dec 01 09:33:33 crc kubenswrapper[4873]: I1201 09:33:33.156132 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76b5fdb995-hdbhs" event={"ID":"90639568-2248-4249-be34-140ed23a5d1d","Type":"ContainerStarted","Data":"f8d8cfb039d747a050fca1d113cb8cb62781c9ecf58e3c77ef2cfefc27923659"} Dec 01 09:33:33 crc kubenswrapper[4873]: I1201 09:33:33.162307 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"8458dfc6-db77-4d48-9c9a-9324813002a3","Type":"ContainerStarted","Data":"a2813e33d30bd5db2a67208cd3dcb1c135568634e26ec74f91f4cfb41aea750e"} Dec 01 09:33:33 crc kubenswrapper[4873]: I1201 09:33:33.162394 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"8458dfc6-db77-4d48-9c9a-9324813002a3","Type":"ContainerStarted","Data":"28a8dcca82c90ec2ca0be16169d4a7e69e98035b1962070e981ac5a9a9d7309c"} Dec 01 09:33:34 crc kubenswrapper[4873]: I1201 09:33:34.142352 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-6968fbf848-w68ch" Dec 01 09:33:34 crc kubenswrapper[4873]: I1201 09:33:34.204091 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76b5fdb995-hdbhs" event={"ID":"90639568-2248-4249-be34-140ed23a5d1d","Type":"ContainerStarted","Data":"1b4758f52cff39e4fb46d2a601aab03c7c19c7d60e3addf5a4d771c4fdffa8ee"} Dec 01 09:33:34 crc kubenswrapper[4873]: I1201 09:33:34.204168 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-76b5fdb995-hdbhs" Dec 01 09:33:34 crc kubenswrapper[4873]: I1201 09:33:34.212347 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"8458dfc6-db77-4d48-9c9a-9324813002a3","Type":"ContainerStarted","Data":"8118b59c0148ab1841e74a0cbd8634349992108260df3f385abc936dd6d542a0"} Dec 01 09:33:34 crc kubenswrapper[4873]: I1201 09:33:34.213507 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Dec 01 09:33:34 crc kubenswrapper[4873]: I1201 09:33:34.217840 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"15f823d3-d00f-4c94-bd11-3cd81c911b10","Type":"ContainerStarted","Data":"8896e391709e5eb0006fc456e6db831869adf7e5269299dbc0f866dfa489dc2a"} Dec 01 09:33:34 crc kubenswrapper[4873]: I1201 09:33:34.239424 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-76b5fdb995-hdbhs" podStartSLOduration=4.239394828 podStartE2EDuration="4.239394828s" podCreationTimestamp="2025-12-01 09:33:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:33:34.229561481 +0000 UTC m=+3190.131670020" watchObservedRunningTime="2025-12-01 09:33:34.239394828 +0000 UTC m=+3190.141503367" Dec 01 09:33:34 crc kubenswrapper[4873]: I1201 09:33:34.277690 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=3.277660967 podStartE2EDuration="3.277660967s" podCreationTimestamp="2025-12-01 09:33:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:33:34.262824475 +0000 UTC m=+3190.164933044" watchObservedRunningTime="2025-12-01 09:33:34.277660967 +0000 UTC m=+3190.179769506" Dec 01 09:33:34 crc kubenswrapper[4873]: I1201 09:33:34.299887 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7c885c5cb8-46fsf" Dec 01 09:33:34 crc kubenswrapper[4873]: I1201 09:33:34.333331 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Dec 01 09:33:35 crc kubenswrapper[4873]: I1201 09:33:35.270922 4873 generic.go:334] "Generic (PLEG): container finished" podID="a963ef99-5387-4db3-bbe3-b3cedcdb899d" containerID="453929b758d99b82c99434bbb65d572a5bf6916ea5c7c47d3ebd6c23242f821c" exitCode=137 Dec 01 09:33:35 crc kubenswrapper[4873]: I1201 09:33:35.271038 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bd6859fdf-xwffw" event={"ID":"a963ef99-5387-4db3-bbe3-b3cedcdb899d","Type":"ContainerDied","Data":"453929b758d99b82c99434bbb65d572a5bf6916ea5c7c47d3ebd6c23242f821c"} Dec 01 09:33:35 crc kubenswrapper[4873]: I1201 09:33:35.285163 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"15f823d3-d00f-4c94-bd11-3cd81c911b10","Type":"ContainerStarted","Data":"a1c44d8a9cc1062f2ff31dc42aacc968640cd8481b9aa168ed813adcdd26fc36"} Dec 01 09:33:35 crc kubenswrapper[4873]: I1201 09:33:35.308357 4873 generic.go:334] "Generic (PLEG): container finished" podID="76d98e93-51f4-49f0-93cd-6564618e6304" containerID="f44ef92bf1b8023c2aa0955521d935b25a93df5d0bb925d1afedb91ba70abe9a" exitCode=137 Dec 01 09:33:35 crc kubenswrapper[4873]: I1201 09:33:35.308413 4873 generic.go:334] "Generic (PLEG): container finished" podID="76d98e93-51f4-49f0-93cd-6564618e6304" containerID="5426af0e98fa5587e33a09bf116d7a4e068bb68eb79d1705830501df2c8917bd" exitCode=137 Dec 01 09:33:35 crc kubenswrapper[4873]: I1201 09:33:35.310067 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66d489fdc7-jqfsv" event={"ID":"76d98e93-51f4-49f0-93cd-6564618e6304","Type":"ContainerDied","Data":"f44ef92bf1b8023c2aa0955521d935b25a93df5d0bb925d1afedb91ba70abe9a"} Dec 01 09:33:35 crc kubenswrapper[4873]: I1201 09:33:35.310112 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66d489fdc7-jqfsv" event={"ID":"76d98e93-51f4-49f0-93cd-6564618e6304","Type":"ContainerDied","Data":"5426af0e98fa5587e33a09bf116d7a4e068bb68eb79d1705830501df2c8917bd"} Dec 01 09:33:35 crc kubenswrapper[4873]: I1201 09:33:35.333305 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=3.920561105 podStartE2EDuration="5.333277542s" podCreationTimestamp="2025-12-01 09:33:30 +0000 UTC" firstStartedPulling="2025-12-01 09:33:31.889140323 +0000 UTC m=+3187.791248862" lastFinishedPulling="2025-12-01 09:33:33.30185676 +0000 UTC m=+3189.203965299" observedRunningTime="2025-12-01 09:33:35.32199366 +0000 UTC m=+3191.224102199" watchObservedRunningTime="2025-12-01 09:33:35.333277542 +0000 UTC m=+3191.235386081" Dec 01 09:33:35 crc kubenswrapper[4873]: I1201 09:33:35.634079 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66d489fdc7-jqfsv" Dec 01 09:33:35 crc kubenswrapper[4873]: I1201 09:33:35.687135 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/76d98e93-51f4-49f0-93cd-6564618e6304-config-data\") pod \"76d98e93-51f4-49f0-93cd-6564618e6304\" (UID: \"76d98e93-51f4-49f0-93cd-6564618e6304\") " Dec 01 09:33:35 crc kubenswrapper[4873]: I1201 09:33:35.687207 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/76d98e93-51f4-49f0-93cd-6564618e6304-scripts\") pod \"76d98e93-51f4-49f0-93cd-6564618e6304\" (UID: \"76d98e93-51f4-49f0-93cd-6564618e6304\") " Dec 01 09:33:35 crc kubenswrapper[4873]: I1201 09:33:35.687387 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/76d98e93-51f4-49f0-93cd-6564618e6304-horizon-secret-key\") pod \"76d98e93-51f4-49f0-93cd-6564618e6304\" (UID: \"76d98e93-51f4-49f0-93cd-6564618e6304\") " Dec 01 09:33:35 crc kubenswrapper[4873]: I1201 09:33:35.687424 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76d98e93-51f4-49f0-93cd-6564618e6304-logs\") pod \"76d98e93-51f4-49f0-93cd-6564618e6304\" (UID: \"76d98e93-51f4-49f0-93cd-6564618e6304\") " Dec 01 09:33:35 crc kubenswrapper[4873]: I1201 09:33:35.687482 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9tpcf\" (UniqueName: \"kubernetes.io/projected/76d98e93-51f4-49f0-93cd-6564618e6304-kube-api-access-9tpcf\") pod \"76d98e93-51f4-49f0-93cd-6564618e6304\" (UID: \"76d98e93-51f4-49f0-93cd-6564618e6304\") " Dec 01 09:33:35 crc kubenswrapper[4873]: I1201 09:33:35.688662 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76d98e93-51f4-49f0-93cd-6564618e6304-logs" (OuterVolumeSpecName: "logs") pod "76d98e93-51f4-49f0-93cd-6564618e6304" (UID: "76d98e93-51f4-49f0-93cd-6564618e6304"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:33:35 crc kubenswrapper[4873]: I1201 09:33:35.699273 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76d98e93-51f4-49f0-93cd-6564618e6304-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "76d98e93-51f4-49f0-93cd-6564618e6304" (UID: "76d98e93-51f4-49f0-93cd-6564618e6304"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:33:35 crc kubenswrapper[4873]: I1201 09:33:35.711516 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76d98e93-51f4-49f0-93cd-6564618e6304-kube-api-access-9tpcf" (OuterVolumeSpecName: "kube-api-access-9tpcf") pod "76d98e93-51f4-49f0-93cd-6564618e6304" (UID: "76d98e93-51f4-49f0-93cd-6564618e6304"). InnerVolumeSpecName "kube-api-access-9tpcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:33:35 crc kubenswrapper[4873]: E1201 09:33:35.787113 4873 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda963ef99_5387_4db3_bbe3_b3cedcdb899d.slice/crio-conmon-6ce94b2693845790bf7cfc2937cf542fcb9a8db1291b9be886142a61b88fd38d.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod76d98e93_51f4_49f0_93cd_6564618e6304.slice/crio-conmon-f44ef92bf1b8023c2aa0955521d935b25a93df5d0bb925d1afedb91ba70abe9a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod76d98e93_51f4_49f0_93cd_6564618e6304.slice/crio-5426af0e98fa5587e33a09bf116d7a4e068bb68eb79d1705830501df2c8917bd.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda963ef99_5387_4db3_bbe3_b3cedcdb899d.slice/crio-6ce94b2693845790bf7cfc2937cf542fcb9a8db1291b9be886142a61b88fd38d.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda963ef99_5387_4db3_bbe3_b3cedcdb899d.slice/crio-453929b758d99b82c99434bbb65d572a5bf6916ea5c7c47d3ebd6c23242f821c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda963ef99_5387_4db3_bbe3_b3cedcdb899d.slice/crio-conmon-453929b758d99b82c99434bbb65d572a5bf6916ea5c7c47d3ebd6c23242f821c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod76d98e93_51f4_49f0_93cd_6564618e6304.slice/crio-f44ef92bf1b8023c2aa0955521d935b25a93df5d0bb925d1afedb91ba70abe9a.scope\": RecentStats: unable to find data in memory cache]" Dec 01 09:33:35 crc kubenswrapper[4873]: I1201 09:33:35.793119 4873 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/76d98e93-51f4-49f0-93cd-6564618e6304-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 01 09:33:35 crc kubenswrapper[4873]: I1201 09:33:35.793903 4873 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76d98e93-51f4-49f0-93cd-6564618e6304-logs\") on node \"crc\" DevicePath \"\"" Dec 01 09:33:35 crc kubenswrapper[4873]: I1201 09:33:35.794106 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9tpcf\" (UniqueName: \"kubernetes.io/projected/76d98e93-51f4-49f0-93cd-6564618e6304-kube-api-access-9tpcf\") on node \"crc\" DevicePath \"\"" Dec 01 09:33:35 crc kubenswrapper[4873]: I1201 09:33:35.798072 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76d98e93-51f4-49f0-93cd-6564618e6304-scripts" (OuterVolumeSpecName: "scripts") pod "76d98e93-51f4-49f0-93cd-6564618e6304" (UID: "76d98e93-51f4-49f0-93cd-6564618e6304"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:33:35 crc kubenswrapper[4873]: I1201 09:33:35.823351 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76d98e93-51f4-49f0-93cd-6564618e6304-config-data" (OuterVolumeSpecName: "config-data") pod "76d98e93-51f4-49f0-93cd-6564618e6304" (UID: "76d98e93-51f4-49f0-93cd-6564618e6304"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:33:35 crc kubenswrapper[4873]: I1201 09:33:35.855783 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7bd6859fdf-xwffw" Dec 01 09:33:35 crc kubenswrapper[4873]: I1201 09:33:35.897913 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a963ef99-5387-4db3-bbe3-b3cedcdb899d-logs\") pod \"a963ef99-5387-4db3-bbe3-b3cedcdb899d\" (UID: \"a963ef99-5387-4db3-bbe3-b3cedcdb899d\") " Dec 01 09:33:35 crc kubenswrapper[4873]: I1201 09:33:35.898370 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a963ef99-5387-4db3-bbe3-b3cedcdb899d-horizon-secret-key\") pod \"a963ef99-5387-4db3-bbe3-b3cedcdb899d\" (UID: \"a963ef99-5387-4db3-bbe3-b3cedcdb899d\") " Dec 01 09:33:35 crc kubenswrapper[4873]: I1201 09:33:35.898484 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a963ef99-5387-4db3-bbe3-b3cedcdb899d-config-data\") pod \"a963ef99-5387-4db3-bbe3-b3cedcdb899d\" (UID: \"a963ef99-5387-4db3-bbe3-b3cedcdb899d\") " Dec 01 09:33:35 crc kubenswrapper[4873]: I1201 09:33:35.898625 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a963ef99-5387-4db3-bbe3-b3cedcdb899d-scripts\") pod \"a963ef99-5387-4db3-bbe3-b3cedcdb899d\" (UID: \"a963ef99-5387-4db3-bbe3-b3cedcdb899d\") " Dec 01 09:33:35 crc kubenswrapper[4873]: I1201 09:33:35.898757 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dzxnb\" (UniqueName: \"kubernetes.io/projected/a963ef99-5387-4db3-bbe3-b3cedcdb899d-kube-api-access-dzxnb\") pod \"a963ef99-5387-4db3-bbe3-b3cedcdb899d\" (UID: \"a963ef99-5387-4db3-bbe3-b3cedcdb899d\") " Dec 01 09:33:35 crc kubenswrapper[4873]: I1201 09:33:35.899453 4873 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/76d98e93-51f4-49f0-93cd-6564618e6304-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:33:35 crc kubenswrapper[4873]: I1201 09:33:35.899520 4873 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/76d98e93-51f4-49f0-93cd-6564618e6304-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:33:35 crc kubenswrapper[4873]: I1201 09:33:35.902421 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a963ef99-5387-4db3-bbe3-b3cedcdb899d-logs" (OuterVolumeSpecName: "logs") pod "a963ef99-5387-4db3-bbe3-b3cedcdb899d" (UID: "a963ef99-5387-4db3-bbe3-b3cedcdb899d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:33:35 crc kubenswrapper[4873]: I1201 09:33:35.905716 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a963ef99-5387-4db3-bbe3-b3cedcdb899d-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "a963ef99-5387-4db3-bbe3-b3cedcdb899d" (UID: "a963ef99-5387-4db3-bbe3-b3cedcdb899d"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:33:35 crc kubenswrapper[4873]: I1201 09:33:35.941840 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a963ef99-5387-4db3-bbe3-b3cedcdb899d-kube-api-access-dzxnb" (OuterVolumeSpecName: "kube-api-access-dzxnb") pod "a963ef99-5387-4db3-bbe3-b3cedcdb899d" (UID: "a963ef99-5387-4db3-bbe3-b3cedcdb899d"). InnerVolumeSpecName "kube-api-access-dzxnb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:33:35 crc kubenswrapper[4873]: I1201 09:33:35.944416 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a963ef99-5387-4db3-bbe3-b3cedcdb899d-scripts" (OuterVolumeSpecName: "scripts") pod "a963ef99-5387-4db3-bbe3-b3cedcdb899d" (UID: "a963ef99-5387-4db3-bbe3-b3cedcdb899d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:33:36 crc kubenswrapper[4873]: I1201 09:33:36.000638 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a963ef99-5387-4db3-bbe3-b3cedcdb899d-config-data" (OuterVolumeSpecName: "config-data") pod "a963ef99-5387-4db3-bbe3-b3cedcdb899d" (UID: "a963ef99-5387-4db3-bbe3-b3cedcdb899d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:33:36 crc kubenswrapper[4873]: I1201 09:33:36.001116 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a963ef99-5387-4db3-bbe3-b3cedcdb899d-config-data\") pod \"a963ef99-5387-4db3-bbe3-b3cedcdb899d\" (UID: \"a963ef99-5387-4db3-bbe3-b3cedcdb899d\") " Dec 01 09:33:36 crc kubenswrapper[4873]: W1201 09:33:36.001173 4873 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/a963ef99-5387-4db3-bbe3-b3cedcdb899d/volumes/kubernetes.io~configmap/config-data Dec 01 09:33:36 crc kubenswrapper[4873]: I1201 09:33:36.001208 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a963ef99-5387-4db3-bbe3-b3cedcdb899d-config-data" (OuterVolumeSpecName: "config-data") pod "a963ef99-5387-4db3-bbe3-b3cedcdb899d" (UID: "a963ef99-5387-4db3-bbe3-b3cedcdb899d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:33:36 crc kubenswrapper[4873]: I1201 09:33:36.002736 4873 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a963ef99-5387-4db3-bbe3-b3cedcdb899d-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 01 09:33:36 crc kubenswrapper[4873]: I1201 09:33:36.002771 4873 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a963ef99-5387-4db3-bbe3-b3cedcdb899d-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:33:36 crc kubenswrapper[4873]: I1201 09:33:36.002786 4873 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a963ef99-5387-4db3-bbe3-b3cedcdb899d-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:33:36 crc kubenswrapper[4873]: I1201 09:33:36.002800 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dzxnb\" (UniqueName: \"kubernetes.io/projected/a963ef99-5387-4db3-bbe3-b3cedcdb899d-kube-api-access-dzxnb\") on node \"crc\" DevicePath \"\"" Dec 01 09:33:36 crc kubenswrapper[4873]: I1201 09:33:36.002817 4873 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a963ef99-5387-4db3-bbe3-b3cedcdb899d-logs\") on node \"crc\" DevicePath \"\"" Dec 01 09:33:36 crc kubenswrapper[4873]: I1201 09:33:36.324586 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66d489fdc7-jqfsv" event={"ID":"76d98e93-51f4-49f0-93cd-6564618e6304","Type":"ContainerDied","Data":"450b355ea6db79f14f51d471535dffa9a11e3538ff03d5dbd44c56063bb965b0"} Dec 01 09:33:36 crc kubenswrapper[4873]: I1201 09:33:36.326045 4873 scope.go:117] "RemoveContainer" containerID="f44ef92bf1b8023c2aa0955521d935b25a93df5d0bb925d1afedb91ba70abe9a" Dec 01 09:33:36 crc kubenswrapper[4873]: I1201 09:33:36.324998 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66d489fdc7-jqfsv" Dec 01 09:33:36 crc kubenswrapper[4873]: I1201 09:33:36.329512 4873 generic.go:334] "Generic (PLEG): container finished" podID="a963ef99-5387-4db3-bbe3-b3cedcdb899d" containerID="6ce94b2693845790bf7cfc2937cf542fcb9a8db1291b9be886142a61b88fd38d" exitCode=137 Dec 01 09:33:36 crc kubenswrapper[4873]: I1201 09:33:36.329888 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7bd6859fdf-xwffw" Dec 01 09:33:36 crc kubenswrapper[4873]: I1201 09:33:36.329878 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bd6859fdf-xwffw" event={"ID":"a963ef99-5387-4db3-bbe3-b3cedcdb899d","Type":"ContainerDied","Data":"6ce94b2693845790bf7cfc2937cf542fcb9a8db1291b9be886142a61b88fd38d"} Dec 01 09:33:36 crc kubenswrapper[4873]: I1201 09:33:36.330297 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bd6859fdf-xwffw" event={"ID":"a963ef99-5387-4db3-bbe3-b3cedcdb899d","Type":"ContainerDied","Data":"6e79504426d789af3cbc7fb0dd6a12879c634e14e55bbe6673b0511f670a845f"} Dec 01 09:33:36 crc kubenswrapper[4873]: I1201 09:33:36.331896 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="8458dfc6-db77-4d48-9c9a-9324813002a3" containerName="manila-api-log" containerID="cri-o://a2813e33d30bd5db2a67208cd3dcb1c135568634e26ec74f91f4cfb41aea750e" gracePeriod=30 Dec 01 09:33:36 crc kubenswrapper[4873]: I1201 09:33:36.331965 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="8458dfc6-db77-4d48-9c9a-9324813002a3" containerName="manila-api" containerID="cri-o://8118b59c0148ab1841e74a0cbd8634349992108260df3f385abc936dd6d542a0" gracePeriod=30 Dec 01 09:33:36 crc kubenswrapper[4873]: I1201 09:33:36.391946 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7bd6859fdf-xwffw"] Dec 01 09:33:36 crc kubenswrapper[4873]: I1201 09:33:36.453507 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7bd6859fdf-xwffw"] Dec 01 09:33:36 crc kubenswrapper[4873]: I1201 09:33:36.454571 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-66d489fdc7-jqfsv"] Dec 01 09:33:36 crc kubenswrapper[4873]: I1201 09:33:36.476303 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-66d489fdc7-jqfsv"] Dec 01 09:33:36 crc kubenswrapper[4873]: I1201 09:33:36.616863 4873 scope.go:117] "RemoveContainer" containerID="5426af0e98fa5587e33a09bf116d7a4e068bb68eb79d1705830501df2c8917bd" Dec 01 09:33:36 crc kubenswrapper[4873]: I1201 09:33:36.771482 4873 scope.go:117] "RemoveContainer" containerID="6ce94b2693845790bf7cfc2937cf542fcb9a8db1291b9be886142a61b88fd38d" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.003409 4873 scope.go:117] "RemoveContainer" containerID="453929b758d99b82c99434bbb65d572a5bf6916ea5c7c47d3ebd6c23242f821c" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.067266 4873 scope.go:117] "RemoveContainer" containerID="6ce94b2693845790bf7cfc2937cf542fcb9a8db1291b9be886142a61b88fd38d" Dec 01 09:33:37 crc kubenswrapper[4873]: E1201 09:33:37.068118 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ce94b2693845790bf7cfc2937cf542fcb9a8db1291b9be886142a61b88fd38d\": container with ID starting with 6ce94b2693845790bf7cfc2937cf542fcb9a8db1291b9be886142a61b88fd38d not found: ID does not exist" containerID="6ce94b2693845790bf7cfc2937cf542fcb9a8db1291b9be886142a61b88fd38d" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.068157 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ce94b2693845790bf7cfc2937cf542fcb9a8db1291b9be886142a61b88fd38d"} err="failed to get container status \"6ce94b2693845790bf7cfc2937cf542fcb9a8db1291b9be886142a61b88fd38d\": rpc error: code = NotFound desc = could not find container \"6ce94b2693845790bf7cfc2937cf542fcb9a8db1291b9be886142a61b88fd38d\": container with ID starting with 6ce94b2693845790bf7cfc2937cf542fcb9a8db1291b9be886142a61b88fd38d not found: ID does not exist" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.068189 4873 scope.go:117] "RemoveContainer" containerID="453929b758d99b82c99434bbb65d572a5bf6916ea5c7c47d3ebd6c23242f821c" Dec 01 09:33:37 crc kubenswrapper[4873]: E1201 09:33:37.071765 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"453929b758d99b82c99434bbb65d572a5bf6916ea5c7c47d3ebd6c23242f821c\": container with ID starting with 453929b758d99b82c99434bbb65d572a5bf6916ea5c7c47d3ebd6c23242f821c not found: ID does not exist" containerID="453929b758d99b82c99434bbb65d572a5bf6916ea5c7c47d3ebd6c23242f821c" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.071836 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"453929b758d99b82c99434bbb65d572a5bf6916ea5c7c47d3ebd6c23242f821c"} err="failed to get container status \"453929b758d99b82c99434bbb65d572a5bf6916ea5c7c47d3ebd6c23242f821c\": rpc error: code = NotFound desc = could not find container \"453929b758d99b82c99434bbb65d572a5bf6916ea5c7c47d3ebd6c23242f821c\": container with ID starting with 453929b758d99b82c99434bbb65d572a5bf6916ea5c7c47d3ebd6c23242f821c not found: ID does not exist" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.249540 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-6968fbf848-w68ch" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.253194 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.324024 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7c885c5cb8-46fsf"] Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.324360 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7c885c5cb8-46fsf" podUID="24452fe2-b990-4537-b619-9db436ed5acb" containerName="horizon-log" containerID="cri-o://3f684c99d8aa2dce24bd09bca0e8fa379cf0d83ddd7621cb5266dddb613a60a8" gracePeriod=30 Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.324899 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7c885c5cb8-46fsf" podUID="24452fe2-b990-4537-b619-9db436ed5acb" containerName="horizon" containerID="cri-o://19a1f4707bbeeffe7274e84c66714fc06a3e1466289970827846d71126cb57c3" gracePeriod=30 Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.347449 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8458dfc6-db77-4d48-9c9a-9324813002a3-etc-machine-id\") pod \"8458dfc6-db77-4d48-9c9a-9324813002a3\" (UID: \"8458dfc6-db77-4d48-9c9a-9324813002a3\") " Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.347515 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8458dfc6-db77-4d48-9c9a-9324813002a3-combined-ca-bundle\") pod \"8458dfc6-db77-4d48-9c9a-9324813002a3\" (UID: \"8458dfc6-db77-4d48-9c9a-9324813002a3\") " Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.347565 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8458dfc6-db77-4d48-9c9a-9324813002a3-config-data\") pod \"8458dfc6-db77-4d48-9c9a-9324813002a3\" (UID: \"8458dfc6-db77-4d48-9c9a-9324813002a3\") " Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.347688 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8458dfc6-db77-4d48-9c9a-9324813002a3-config-data-custom\") pod \"8458dfc6-db77-4d48-9c9a-9324813002a3\" (UID: \"8458dfc6-db77-4d48-9c9a-9324813002a3\") " Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.347778 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8458dfc6-db77-4d48-9c9a-9324813002a3-scripts\") pod \"8458dfc6-db77-4d48-9c9a-9324813002a3\" (UID: \"8458dfc6-db77-4d48-9c9a-9324813002a3\") " Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.347833 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s6zh7\" (UniqueName: \"kubernetes.io/projected/8458dfc6-db77-4d48-9c9a-9324813002a3-kube-api-access-s6zh7\") pod \"8458dfc6-db77-4d48-9c9a-9324813002a3\" (UID: \"8458dfc6-db77-4d48-9c9a-9324813002a3\") " Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.348071 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8458dfc6-db77-4d48-9c9a-9324813002a3-logs\") pod \"8458dfc6-db77-4d48-9c9a-9324813002a3\" (UID: \"8458dfc6-db77-4d48-9c9a-9324813002a3\") " Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.363432 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8458dfc6-db77-4d48-9c9a-9324813002a3-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "8458dfc6-db77-4d48-9c9a-9324813002a3" (UID: "8458dfc6-db77-4d48-9c9a-9324813002a3"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.380275 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8458dfc6-db77-4d48-9c9a-9324813002a3-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "8458dfc6-db77-4d48-9c9a-9324813002a3" (UID: "8458dfc6-db77-4d48-9c9a-9324813002a3"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.380288 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8458dfc6-db77-4d48-9c9a-9324813002a3-scripts" (OuterVolumeSpecName: "scripts") pod "8458dfc6-db77-4d48-9c9a-9324813002a3" (UID: "8458dfc6-db77-4d48-9c9a-9324813002a3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.380894 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8458dfc6-db77-4d48-9c9a-9324813002a3-logs" (OuterVolumeSpecName: "logs") pod "8458dfc6-db77-4d48-9c9a-9324813002a3" (UID: "8458dfc6-db77-4d48-9c9a-9324813002a3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.387457 4873 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7c885c5cb8-46fsf" podUID="24452fe2-b990-4537-b619-9db436ed5acb" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.242:8443/dashboard/auth/login/?next=/dashboard/\": EOF" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.394270 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8458dfc6-db77-4d48-9c9a-9324813002a3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8458dfc6-db77-4d48-9c9a-9324813002a3" (UID: "8458dfc6-db77-4d48-9c9a-9324813002a3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.417320 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8458dfc6-db77-4d48-9c9a-9324813002a3-kube-api-access-s6zh7" (OuterVolumeSpecName: "kube-api-access-s6zh7") pod "8458dfc6-db77-4d48-9c9a-9324813002a3" (UID: "8458dfc6-db77-4d48-9c9a-9324813002a3"). InnerVolumeSpecName "kube-api-access-s6zh7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.436633 4873 scope.go:117] "RemoveContainer" containerID="1eb6abed9ec75ad80889e13d075ac04aa13db51de45fa2f980efa98223444319" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.441849 4873 generic.go:334] "Generic (PLEG): container finished" podID="8458dfc6-db77-4d48-9c9a-9324813002a3" containerID="8118b59c0148ab1841e74a0cbd8634349992108260df3f385abc936dd6d542a0" exitCode=0 Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.441908 4873 generic.go:334] "Generic (PLEG): container finished" podID="8458dfc6-db77-4d48-9c9a-9324813002a3" containerID="a2813e33d30bd5db2a67208cd3dcb1c135568634e26ec74f91f4cfb41aea750e" exitCode=143 Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.442002 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"8458dfc6-db77-4d48-9c9a-9324813002a3","Type":"ContainerDied","Data":"8118b59c0148ab1841e74a0cbd8634349992108260df3f385abc936dd6d542a0"} Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.442063 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"8458dfc6-db77-4d48-9c9a-9324813002a3","Type":"ContainerDied","Data":"a2813e33d30bd5db2a67208cd3dcb1c135568634e26ec74f91f4cfb41aea750e"} Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.442077 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"8458dfc6-db77-4d48-9c9a-9324813002a3","Type":"ContainerDied","Data":"28a8dcca82c90ec2ca0be16169d4a7e69e98035b1962070e981ac5a9a9d7309c"} Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.442099 4873 scope.go:117] "RemoveContainer" containerID="8118b59c0148ab1841e74a0cbd8634349992108260df3f385abc936dd6d542a0" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.442291 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.466390 4873 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8458dfc6-db77-4d48-9c9a-9324813002a3-logs\") on node \"crc\" DevicePath \"\"" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.466429 4873 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8458dfc6-db77-4d48-9c9a-9324813002a3-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.466444 4873 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8458dfc6-db77-4d48-9c9a-9324813002a3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.466456 4873 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8458dfc6-db77-4d48-9c9a-9324813002a3-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.466466 4873 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8458dfc6-db77-4d48-9c9a-9324813002a3-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.466479 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s6zh7\" (UniqueName: \"kubernetes.io/projected/8458dfc6-db77-4d48-9c9a-9324813002a3-kube-api-access-s6zh7\") on node \"crc\" DevicePath \"\"" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.503749 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8458dfc6-db77-4d48-9c9a-9324813002a3-config-data" (OuterVolumeSpecName: "config-data") pod "8458dfc6-db77-4d48-9c9a-9324813002a3" (UID: "8458dfc6-db77-4d48-9c9a-9324813002a3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.570542 4873 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8458dfc6-db77-4d48-9c9a-9324813002a3-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.576430 4873 scope.go:117] "RemoveContainer" containerID="a2813e33d30bd5db2a67208cd3dcb1c135568634e26ec74f91f4cfb41aea750e" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.616454 4873 scope.go:117] "RemoveContainer" containerID="8118b59c0148ab1841e74a0cbd8634349992108260df3f385abc936dd6d542a0" Dec 01 09:33:37 crc kubenswrapper[4873]: E1201 09:33:37.617759 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8118b59c0148ab1841e74a0cbd8634349992108260df3f385abc936dd6d542a0\": container with ID starting with 8118b59c0148ab1841e74a0cbd8634349992108260df3f385abc936dd6d542a0 not found: ID does not exist" containerID="8118b59c0148ab1841e74a0cbd8634349992108260df3f385abc936dd6d542a0" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.617927 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8118b59c0148ab1841e74a0cbd8634349992108260df3f385abc936dd6d542a0"} err="failed to get container status \"8118b59c0148ab1841e74a0cbd8634349992108260df3f385abc936dd6d542a0\": rpc error: code = NotFound desc = could not find container \"8118b59c0148ab1841e74a0cbd8634349992108260df3f385abc936dd6d542a0\": container with ID starting with 8118b59c0148ab1841e74a0cbd8634349992108260df3f385abc936dd6d542a0 not found: ID does not exist" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.617957 4873 scope.go:117] "RemoveContainer" containerID="a2813e33d30bd5db2a67208cd3dcb1c135568634e26ec74f91f4cfb41aea750e" Dec 01 09:33:37 crc kubenswrapper[4873]: E1201 09:33:37.619469 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2813e33d30bd5db2a67208cd3dcb1c135568634e26ec74f91f4cfb41aea750e\": container with ID starting with a2813e33d30bd5db2a67208cd3dcb1c135568634e26ec74f91f4cfb41aea750e not found: ID does not exist" containerID="a2813e33d30bd5db2a67208cd3dcb1c135568634e26ec74f91f4cfb41aea750e" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.619587 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2813e33d30bd5db2a67208cd3dcb1c135568634e26ec74f91f4cfb41aea750e"} err="failed to get container status \"a2813e33d30bd5db2a67208cd3dcb1c135568634e26ec74f91f4cfb41aea750e\": rpc error: code = NotFound desc = could not find container \"a2813e33d30bd5db2a67208cd3dcb1c135568634e26ec74f91f4cfb41aea750e\": container with ID starting with a2813e33d30bd5db2a67208cd3dcb1c135568634e26ec74f91f4cfb41aea750e not found: ID does not exist" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.619609 4873 scope.go:117] "RemoveContainer" containerID="8118b59c0148ab1841e74a0cbd8634349992108260df3f385abc936dd6d542a0" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.619933 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8118b59c0148ab1841e74a0cbd8634349992108260df3f385abc936dd6d542a0"} err="failed to get container status \"8118b59c0148ab1841e74a0cbd8634349992108260df3f385abc936dd6d542a0\": rpc error: code = NotFound desc = could not find container \"8118b59c0148ab1841e74a0cbd8634349992108260df3f385abc936dd6d542a0\": container with ID starting with 8118b59c0148ab1841e74a0cbd8634349992108260df3f385abc936dd6d542a0 not found: ID does not exist" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.619957 4873 scope.go:117] "RemoveContainer" containerID="a2813e33d30bd5db2a67208cd3dcb1c135568634e26ec74f91f4cfb41aea750e" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.622119 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2813e33d30bd5db2a67208cd3dcb1c135568634e26ec74f91f4cfb41aea750e"} err="failed to get container status \"a2813e33d30bd5db2a67208cd3dcb1c135568634e26ec74f91f4cfb41aea750e\": rpc error: code = NotFound desc = could not find container \"a2813e33d30bd5db2a67208cd3dcb1c135568634e26ec74f91f4cfb41aea750e\": container with ID starting with a2813e33d30bd5db2a67208cd3dcb1c135568634e26ec74f91f4cfb41aea750e not found: ID does not exist" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.799228 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.830247 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-api-0"] Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.853190 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Dec 01 09:33:37 crc kubenswrapper[4873]: E1201 09:33:37.853737 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76d98e93-51f4-49f0-93cd-6564618e6304" containerName="horizon" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.853758 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="76d98e93-51f4-49f0-93cd-6564618e6304" containerName="horizon" Dec 01 09:33:37 crc kubenswrapper[4873]: E1201 09:33:37.853769 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8458dfc6-db77-4d48-9c9a-9324813002a3" containerName="manila-api" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.853777 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="8458dfc6-db77-4d48-9c9a-9324813002a3" containerName="manila-api" Dec 01 09:33:37 crc kubenswrapper[4873]: E1201 09:33:37.853791 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76d98e93-51f4-49f0-93cd-6564618e6304" containerName="horizon-log" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.853798 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="76d98e93-51f4-49f0-93cd-6564618e6304" containerName="horizon-log" Dec 01 09:33:37 crc kubenswrapper[4873]: E1201 09:33:37.853810 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8458dfc6-db77-4d48-9c9a-9324813002a3" containerName="manila-api-log" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.853816 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="8458dfc6-db77-4d48-9c9a-9324813002a3" containerName="manila-api-log" Dec 01 09:33:37 crc kubenswrapper[4873]: E1201 09:33:37.853831 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a963ef99-5387-4db3-bbe3-b3cedcdb899d" containerName="horizon-log" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.853837 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="a963ef99-5387-4db3-bbe3-b3cedcdb899d" containerName="horizon-log" Dec 01 09:33:37 crc kubenswrapper[4873]: E1201 09:33:37.853861 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a963ef99-5387-4db3-bbe3-b3cedcdb899d" containerName="horizon" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.853869 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="a963ef99-5387-4db3-bbe3-b3cedcdb899d" containerName="horizon" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.854346 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="a963ef99-5387-4db3-bbe3-b3cedcdb899d" containerName="horizon" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.854373 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="a963ef99-5387-4db3-bbe3-b3cedcdb899d" containerName="horizon-log" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.854392 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="76d98e93-51f4-49f0-93cd-6564618e6304" containerName="horizon-log" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.854405 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="8458dfc6-db77-4d48-9c9a-9324813002a3" containerName="manila-api" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.854416 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="76d98e93-51f4-49f0-93cd-6564618e6304" containerName="horizon" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.854425 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="8458dfc6-db77-4d48-9c9a-9324813002a3" containerName="manila-api-log" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.855703 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.865469 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-internal-svc" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.868959 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-public-svc" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.872075 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.902154 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.983912 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7744a110-0501-4606-9c09-a6442c0fefb0-logs\") pod \"manila-api-0\" (UID: \"7744a110-0501-4606-9c09-a6442c0fefb0\") " pod="openstack/manila-api-0" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.984432 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7744a110-0501-4606-9c09-a6442c0fefb0-etc-machine-id\") pod \"manila-api-0\" (UID: \"7744a110-0501-4606-9c09-a6442c0fefb0\") " pod="openstack/manila-api-0" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.984619 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfqf8\" (UniqueName: \"kubernetes.io/projected/7744a110-0501-4606-9c09-a6442c0fefb0-kube-api-access-xfqf8\") pod \"manila-api-0\" (UID: \"7744a110-0501-4606-9c09-a6442c0fefb0\") " pod="openstack/manila-api-0" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.984679 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7744a110-0501-4606-9c09-a6442c0fefb0-internal-tls-certs\") pod \"manila-api-0\" (UID: \"7744a110-0501-4606-9c09-a6442c0fefb0\") " pod="openstack/manila-api-0" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.984711 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7744a110-0501-4606-9c09-a6442c0fefb0-config-data\") pod \"manila-api-0\" (UID: \"7744a110-0501-4606-9c09-a6442c0fefb0\") " pod="openstack/manila-api-0" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.984783 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7744a110-0501-4606-9c09-a6442c0fefb0-scripts\") pod \"manila-api-0\" (UID: \"7744a110-0501-4606-9c09-a6442c0fefb0\") " pod="openstack/manila-api-0" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.984814 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7744a110-0501-4606-9c09-a6442c0fefb0-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"7744a110-0501-4606-9c09-a6442c0fefb0\") " pod="openstack/manila-api-0" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.984833 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7744a110-0501-4606-9c09-a6442c0fefb0-config-data-custom\") pod \"manila-api-0\" (UID: \"7744a110-0501-4606-9c09-a6442c0fefb0\") " pod="openstack/manila-api-0" Dec 01 09:33:37 crc kubenswrapper[4873]: I1201 09:33:37.984872 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7744a110-0501-4606-9c09-a6442c0fefb0-public-tls-certs\") pod \"manila-api-0\" (UID: \"7744a110-0501-4606-9c09-a6442c0fefb0\") " pod="openstack/manila-api-0" Dec 01 09:33:38 crc kubenswrapper[4873]: I1201 09:33:38.085589 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfqf8\" (UniqueName: \"kubernetes.io/projected/7744a110-0501-4606-9c09-a6442c0fefb0-kube-api-access-xfqf8\") pod \"manila-api-0\" (UID: \"7744a110-0501-4606-9c09-a6442c0fefb0\") " pod="openstack/manila-api-0" Dec 01 09:33:38 crc kubenswrapper[4873]: I1201 09:33:38.086068 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7744a110-0501-4606-9c09-a6442c0fefb0-internal-tls-certs\") pod \"manila-api-0\" (UID: \"7744a110-0501-4606-9c09-a6442c0fefb0\") " pod="openstack/manila-api-0" Dec 01 09:33:38 crc kubenswrapper[4873]: I1201 09:33:38.086105 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7744a110-0501-4606-9c09-a6442c0fefb0-config-data\") pod \"manila-api-0\" (UID: \"7744a110-0501-4606-9c09-a6442c0fefb0\") " pod="openstack/manila-api-0" Dec 01 09:33:38 crc kubenswrapper[4873]: I1201 09:33:38.086152 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7744a110-0501-4606-9c09-a6442c0fefb0-scripts\") pod \"manila-api-0\" (UID: \"7744a110-0501-4606-9c09-a6442c0fefb0\") " pod="openstack/manila-api-0" Dec 01 09:33:38 crc kubenswrapper[4873]: I1201 09:33:38.086178 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7744a110-0501-4606-9c09-a6442c0fefb0-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"7744a110-0501-4606-9c09-a6442c0fefb0\") " pod="openstack/manila-api-0" Dec 01 09:33:38 crc kubenswrapper[4873]: I1201 09:33:38.086208 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7744a110-0501-4606-9c09-a6442c0fefb0-config-data-custom\") pod \"manila-api-0\" (UID: \"7744a110-0501-4606-9c09-a6442c0fefb0\") " pod="openstack/manila-api-0" Dec 01 09:33:38 crc kubenswrapper[4873]: I1201 09:33:38.086235 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7744a110-0501-4606-9c09-a6442c0fefb0-public-tls-certs\") pod \"manila-api-0\" (UID: \"7744a110-0501-4606-9c09-a6442c0fefb0\") " pod="openstack/manila-api-0" Dec 01 09:33:38 crc kubenswrapper[4873]: I1201 09:33:38.086265 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7744a110-0501-4606-9c09-a6442c0fefb0-logs\") pod \"manila-api-0\" (UID: \"7744a110-0501-4606-9c09-a6442c0fefb0\") " pod="openstack/manila-api-0" Dec 01 09:33:38 crc kubenswrapper[4873]: I1201 09:33:38.086370 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7744a110-0501-4606-9c09-a6442c0fefb0-etc-machine-id\") pod \"manila-api-0\" (UID: \"7744a110-0501-4606-9c09-a6442c0fefb0\") " pod="openstack/manila-api-0" Dec 01 09:33:38 crc kubenswrapper[4873]: I1201 09:33:38.086534 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7744a110-0501-4606-9c09-a6442c0fefb0-etc-machine-id\") pod \"manila-api-0\" (UID: \"7744a110-0501-4606-9c09-a6442c0fefb0\") " pod="openstack/manila-api-0" Dec 01 09:33:38 crc kubenswrapper[4873]: I1201 09:33:38.087656 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7744a110-0501-4606-9c09-a6442c0fefb0-logs\") pod \"manila-api-0\" (UID: \"7744a110-0501-4606-9c09-a6442c0fefb0\") " pod="openstack/manila-api-0" Dec 01 09:33:38 crc kubenswrapper[4873]: I1201 09:33:38.095551 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7744a110-0501-4606-9c09-a6442c0fefb0-config-data-custom\") pod \"manila-api-0\" (UID: \"7744a110-0501-4606-9c09-a6442c0fefb0\") " pod="openstack/manila-api-0" Dec 01 09:33:38 crc kubenswrapper[4873]: I1201 09:33:38.095920 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7744a110-0501-4606-9c09-a6442c0fefb0-internal-tls-certs\") pod \"manila-api-0\" (UID: \"7744a110-0501-4606-9c09-a6442c0fefb0\") " pod="openstack/manila-api-0" Dec 01 09:33:38 crc kubenswrapper[4873]: I1201 09:33:38.095960 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7744a110-0501-4606-9c09-a6442c0fefb0-scripts\") pod \"manila-api-0\" (UID: \"7744a110-0501-4606-9c09-a6442c0fefb0\") " pod="openstack/manila-api-0" Dec 01 09:33:38 crc kubenswrapper[4873]: I1201 09:33:38.109674 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7744a110-0501-4606-9c09-a6442c0fefb0-public-tls-certs\") pod \"manila-api-0\" (UID: \"7744a110-0501-4606-9c09-a6442c0fefb0\") " pod="openstack/manila-api-0" Dec 01 09:33:38 crc kubenswrapper[4873]: I1201 09:33:38.112894 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7744a110-0501-4606-9c09-a6442c0fefb0-config-data\") pod \"manila-api-0\" (UID: \"7744a110-0501-4606-9c09-a6442c0fefb0\") " pod="openstack/manila-api-0" Dec 01 09:33:38 crc kubenswrapper[4873]: I1201 09:33:38.113854 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfqf8\" (UniqueName: \"kubernetes.io/projected/7744a110-0501-4606-9c09-a6442c0fefb0-kube-api-access-xfqf8\") pod \"manila-api-0\" (UID: \"7744a110-0501-4606-9c09-a6442c0fefb0\") " pod="openstack/manila-api-0" Dec 01 09:33:38 crc kubenswrapper[4873]: I1201 09:33:38.116886 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7744a110-0501-4606-9c09-a6442c0fefb0-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"7744a110-0501-4606-9c09-a6442c0fefb0\") " pod="openstack/manila-api-0" Dec 01 09:33:38 crc kubenswrapper[4873]: I1201 09:33:38.182618 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 01 09:33:38 crc kubenswrapper[4873]: I1201 09:33:38.505624 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76d98e93-51f4-49f0-93cd-6564618e6304" path="/var/lib/kubelet/pods/76d98e93-51f4-49f0-93cd-6564618e6304/volumes" Dec 01 09:33:38 crc kubenswrapper[4873]: I1201 09:33:38.506919 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8458dfc6-db77-4d48-9c9a-9324813002a3" path="/var/lib/kubelet/pods/8458dfc6-db77-4d48-9c9a-9324813002a3/volumes" Dec 01 09:33:38 crc kubenswrapper[4873]: I1201 09:33:38.508005 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a963ef99-5387-4db3-bbe3-b3cedcdb899d" path="/var/lib/kubelet/pods/a963ef99-5387-4db3-bbe3-b3cedcdb899d/volumes" Dec 01 09:33:38 crc kubenswrapper[4873]: I1201 09:33:38.617946 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" event={"ID":"fef7b114-0e07-402d-a37b-315c36011f4b","Type":"ContainerStarted","Data":"0b72d3c91d2960798896e9520de55a4bb91b7db511fab82dd6f7e66083a742f2"} Dec 01 09:33:38 crc kubenswrapper[4873]: I1201 09:33:38.940856 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Dec 01 09:33:38 crc kubenswrapper[4873]: W1201 09:33:38.957154 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7744a110_0501_4606_9c09_a6442c0fefb0.slice/crio-b5131ed6910853f0979ca3c26d0913acd2a7b2e744441cae0a57681494cfda19 WatchSource:0}: Error finding container b5131ed6910853f0979ca3c26d0913acd2a7b2e744441cae0a57681494cfda19: Status 404 returned error can't find the container with id b5131ed6910853f0979ca3c26d0913acd2a7b2e744441cae0a57681494cfda19 Dec 01 09:33:39 crc kubenswrapper[4873]: I1201 09:33:39.638011 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"7744a110-0501-4606-9c09-a6442c0fefb0","Type":"ContainerStarted","Data":"0af9501e57c150267bd364aa1d34370712d0fdaf28d091bfee7f66fa691ee0ea"} Dec 01 09:33:39 crc kubenswrapper[4873]: I1201 09:33:39.639154 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"7744a110-0501-4606-9c09-a6442c0fefb0","Type":"ContainerStarted","Data":"b5131ed6910853f0979ca3c26d0913acd2a7b2e744441cae0a57681494cfda19"} Dec 01 09:33:40 crc kubenswrapper[4873]: I1201 09:33:40.652779 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"7744a110-0501-4606-9c09-a6442c0fefb0","Type":"ContainerStarted","Data":"d149ca434e479ed4963848e77dd7f866837d3f1fe132f75454d35503a0023c26"} Dec 01 09:33:40 crc kubenswrapper[4873]: I1201 09:33:40.653364 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Dec 01 09:33:40 crc kubenswrapper[4873]: I1201 09:33:40.686655 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=3.686603501 podStartE2EDuration="3.686603501s" podCreationTimestamp="2025-12-01 09:33:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:33:40.67579947 +0000 UTC m=+3196.577908009" watchObservedRunningTime="2025-12-01 09:33:40.686603501 +0000 UTC m=+3196.588712040" Dec 01 09:33:40 crc kubenswrapper[4873]: I1201 09:33:40.832821 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Dec 01 09:33:41 crc kubenswrapper[4873]: I1201 09:33:41.389817 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-76b5fdb995-hdbhs" Dec 01 09:33:41 crc kubenswrapper[4873]: I1201 09:33:41.480704 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-864d5fc68c-r9f7f"] Dec 01 09:33:41 crc kubenswrapper[4873]: I1201 09:33:41.481100 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-864d5fc68c-r9f7f" podUID="8d7e3564-5684-4792-b3b0-57001892f591" containerName="dnsmasq-dns" containerID="cri-o://26f55973bcf4d0bcc3ac93c0eb1796034503b90c45a1929dc634866c66402344" gracePeriod=10 Dec 01 09:33:41 crc kubenswrapper[4873]: I1201 09:33:41.679252 4873 generic.go:334] "Generic (PLEG): container finished" podID="8d7e3564-5684-4792-b3b0-57001892f591" containerID="26f55973bcf4d0bcc3ac93c0eb1796034503b90c45a1929dc634866c66402344" exitCode=0 Dec 01 09:33:41 crc kubenswrapper[4873]: I1201 09:33:41.679502 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-864d5fc68c-r9f7f" event={"ID":"8d7e3564-5684-4792-b3b0-57001892f591","Type":"ContainerDied","Data":"26f55973bcf4d0bcc3ac93c0eb1796034503b90c45a1929dc634866c66402344"} Dec 01 09:33:41 crc kubenswrapper[4873]: I1201 09:33:41.893454 4873 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7c885c5cb8-46fsf" podUID="24452fe2-b990-4537-b619-9db436ed5acb" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.242:8443/dashboard/auth/login/?next=/dashboard/\": read tcp 10.217.0.2:55254->10.217.0.242:8443: read: connection reset by peer" Dec 01 09:33:41 crc kubenswrapper[4873]: I1201 09:33:41.894801 4873 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7c885c5cb8-46fsf" podUID="24452fe2-b990-4537-b619-9db436ed5acb" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.242:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.242:8443: connect: connection refused" Dec 01 09:33:42 crc kubenswrapper[4873]: I1201 09:33:42.170483 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-864d5fc68c-r9f7f" Dec 01 09:33:42 crc kubenswrapper[4873]: I1201 09:33:42.351075 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8d7e3564-5684-4792-b3b0-57001892f591-ovsdbserver-sb\") pod \"8d7e3564-5684-4792-b3b0-57001892f591\" (UID: \"8d7e3564-5684-4792-b3b0-57001892f591\") " Dec 01 09:33:42 crc kubenswrapper[4873]: I1201 09:33:42.351197 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vcc8x\" (UniqueName: \"kubernetes.io/projected/8d7e3564-5684-4792-b3b0-57001892f591-kube-api-access-vcc8x\") pod \"8d7e3564-5684-4792-b3b0-57001892f591\" (UID: \"8d7e3564-5684-4792-b3b0-57001892f591\") " Dec 01 09:33:42 crc kubenswrapper[4873]: I1201 09:33:42.351260 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/8d7e3564-5684-4792-b3b0-57001892f591-openstack-edpm-ipam\") pod \"8d7e3564-5684-4792-b3b0-57001892f591\" (UID: \"8d7e3564-5684-4792-b3b0-57001892f591\") " Dec 01 09:33:42 crc kubenswrapper[4873]: I1201 09:33:42.351341 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8d7e3564-5684-4792-b3b0-57001892f591-dns-svc\") pod \"8d7e3564-5684-4792-b3b0-57001892f591\" (UID: \"8d7e3564-5684-4792-b3b0-57001892f591\") " Dec 01 09:33:42 crc kubenswrapper[4873]: I1201 09:33:42.351382 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d7e3564-5684-4792-b3b0-57001892f591-config\") pod \"8d7e3564-5684-4792-b3b0-57001892f591\" (UID: \"8d7e3564-5684-4792-b3b0-57001892f591\") " Dec 01 09:33:42 crc kubenswrapper[4873]: I1201 09:33:42.351550 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8d7e3564-5684-4792-b3b0-57001892f591-ovsdbserver-nb\") pod \"8d7e3564-5684-4792-b3b0-57001892f591\" (UID: \"8d7e3564-5684-4792-b3b0-57001892f591\") " Dec 01 09:33:42 crc kubenswrapper[4873]: I1201 09:33:42.398429 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d7e3564-5684-4792-b3b0-57001892f591-kube-api-access-vcc8x" (OuterVolumeSpecName: "kube-api-access-vcc8x") pod "8d7e3564-5684-4792-b3b0-57001892f591" (UID: "8d7e3564-5684-4792-b3b0-57001892f591"). InnerVolumeSpecName "kube-api-access-vcc8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:33:42 crc kubenswrapper[4873]: I1201 09:33:42.471840 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vcc8x\" (UniqueName: \"kubernetes.io/projected/8d7e3564-5684-4792-b3b0-57001892f591-kube-api-access-vcc8x\") on node \"crc\" DevicePath \"\"" Dec 01 09:33:42 crc kubenswrapper[4873]: I1201 09:33:42.536182 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d7e3564-5684-4792-b3b0-57001892f591-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8d7e3564-5684-4792-b3b0-57001892f591" (UID: "8d7e3564-5684-4792-b3b0-57001892f591"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:33:42 crc kubenswrapper[4873]: I1201 09:33:42.566900 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d7e3564-5684-4792-b3b0-57001892f591-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8d7e3564-5684-4792-b3b0-57001892f591" (UID: "8d7e3564-5684-4792-b3b0-57001892f591"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:33:42 crc kubenswrapper[4873]: I1201 09:33:42.575430 4873 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8d7e3564-5684-4792-b3b0-57001892f591-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 09:33:42 crc kubenswrapper[4873]: I1201 09:33:42.576198 4873 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8d7e3564-5684-4792-b3b0-57001892f591-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 09:33:42 crc kubenswrapper[4873]: I1201 09:33:42.585508 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d7e3564-5684-4792-b3b0-57001892f591-config" (OuterVolumeSpecName: "config") pod "8d7e3564-5684-4792-b3b0-57001892f591" (UID: "8d7e3564-5684-4792-b3b0-57001892f591"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:33:42 crc kubenswrapper[4873]: I1201 09:33:42.588407 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d7e3564-5684-4792-b3b0-57001892f591-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "8d7e3564-5684-4792-b3b0-57001892f591" (UID: "8d7e3564-5684-4792-b3b0-57001892f591"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:33:42 crc kubenswrapper[4873]: I1201 09:33:42.590085 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d7e3564-5684-4792-b3b0-57001892f591-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8d7e3564-5684-4792-b3b0-57001892f591" (UID: "8d7e3564-5684-4792-b3b0-57001892f591"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:33:42 crc kubenswrapper[4873]: I1201 09:33:42.678185 4873 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8d7e3564-5684-4792-b3b0-57001892f591-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 09:33:42 crc kubenswrapper[4873]: I1201 09:33:42.678224 4873 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/8d7e3564-5684-4792-b3b0-57001892f591-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 01 09:33:42 crc kubenswrapper[4873]: I1201 09:33:42.678236 4873 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d7e3564-5684-4792-b3b0-57001892f591-config\") on node \"crc\" DevicePath \"\"" Dec 01 09:33:42 crc kubenswrapper[4873]: I1201 09:33:42.695900 4873 generic.go:334] "Generic (PLEG): container finished" podID="24452fe2-b990-4537-b619-9db436ed5acb" containerID="19a1f4707bbeeffe7274e84c66714fc06a3e1466289970827846d71126cb57c3" exitCode=0 Dec 01 09:33:42 crc kubenswrapper[4873]: I1201 09:33:42.695986 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c885c5cb8-46fsf" event={"ID":"24452fe2-b990-4537-b619-9db436ed5acb","Type":"ContainerDied","Data":"19a1f4707bbeeffe7274e84c66714fc06a3e1466289970827846d71126cb57c3"} Dec 01 09:33:42 crc kubenswrapper[4873]: I1201 09:33:42.699871 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-864d5fc68c-r9f7f" event={"ID":"8d7e3564-5684-4792-b3b0-57001892f591","Type":"ContainerDied","Data":"63c28c34179fa261df3a34e7fcbb87902a2b62dab87592559d5813b491e11307"} Dec 01 09:33:42 crc kubenswrapper[4873]: I1201 09:33:42.699950 4873 scope.go:117] "RemoveContainer" containerID="26f55973bcf4d0bcc3ac93c0eb1796034503b90c45a1929dc634866c66402344" Dec 01 09:33:42 crc kubenswrapper[4873]: I1201 09:33:42.699947 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-864d5fc68c-r9f7f" Dec 01 09:33:42 crc kubenswrapper[4873]: I1201 09:33:42.754754 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-864d5fc68c-r9f7f"] Dec 01 09:33:42 crc kubenswrapper[4873]: I1201 09:33:42.768176 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-864d5fc68c-r9f7f"] Dec 01 09:33:44 crc kubenswrapper[4873]: I1201 09:33:44.450860 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d7e3564-5684-4792-b3b0-57001892f591" path="/var/lib/kubelet/pods/8d7e3564-5684-4792-b3b0-57001892f591/volumes" Dec 01 09:33:44 crc kubenswrapper[4873]: I1201 09:33:44.952946 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 09:33:44 crc kubenswrapper[4873]: I1201 09:33:44.953413 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c7cad829-b344-4e49-b0d4-264750c0ae0f" containerName="ceilometer-central-agent" containerID="cri-o://f168c0536d36e29586d04755ace417b659bf6cc437bea36b669ff4b731d89e24" gracePeriod=30 Dec 01 09:33:44 crc kubenswrapper[4873]: I1201 09:33:44.953481 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c7cad829-b344-4e49-b0d4-264750c0ae0f" containerName="proxy-httpd" containerID="cri-o://2bd108f1fad71742ecc62b281d05995aefb751321590b194fa66df222271b3eb" gracePeriod=30 Dec 01 09:33:44 crc kubenswrapper[4873]: I1201 09:33:44.953520 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c7cad829-b344-4e49-b0d4-264750c0ae0f" containerName="sg-core" containerID="cri-o://c91515ebd87bb1e28b9586fe19335671f8d956b8ecdad460712529b7d0846130" gracePeriod=30 Dec 01 09:33:44 crc kubenswrapper[4873]: I1201 09:33:44.953565 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c7cad829-b344-4e49-b0d4-264750c0ae0f" containerName="ceilometer-notification-agent" containerID="cri-o://515feff768ee635a4bc61549e094ad2a0d6362338a61af5ceeedf238ceb452c4" gracePeriod=30 Dec 01 09:33:45 crc kubenswrapper[4873]: I1201 09:33:45.751577 4873 generic.go:334] "Generic (PLEG): container finished" podID="c7cad829-b344-4e49-b0d4-264750c0ae0f" containerID="2bd108f1fad71742ecc62b281d05995aefb751321590b194fa66df222271b3eb" exitCode=0 Dec 01 09:33:45 crc kubenswrapper[4873]: I1201 09:33:45.751914 4873 generic.go:334] "Generic (PLEG): container finished" podID="c7cad829-b344-4e49-b0d4-264750c0ae0f" containerID="c91515ebd87bb1e28b9586fe19335671f8d956b8ecdad460712529b7d0846130" exitCode=2 Dec 01 09:33:45 crc kubenswrapper[4873]: I1201 09:33:45.751926 4873 generic.go:334] "Generic (PLEG): container finished" podID="c7cad829-b344-4e49-b0d4-264750c0ae0f" containerID="f168c0536d36e29586d04755ace417b659bf6cc437bea36b669ff4b731d89e24" exitCode=0 Dec 01 09:33:45 crc kubenswrapper[4873]: I1201 09:33:45.751794 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7cad829-b344-4e49-b0d4-264750c0ae0f","Type":"ContainerDied","Data":"2bd108f1fad71742ecc62b281d05995aefb751321590b194fa66df222271b3eb"} Dec 01 09:33:45 crc kubenswrapper[4873]: I1201 09:33:45.751971 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7cad829-b344-4e49-b0d4-264750c0ae0f","Type":"ContainerDied","Data":"c91515ebd87bb1e28b9586fe19335671f8d956b8ecdad460712529b7d0846130"} Dec 01 09:33:45 crc kubenswrapper[4873]: I1201 09:33:45.751988 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7cad829-b344-4e49-b0d4-264750c0ae0f","Type":"ContainerDied","Data":"f168c0536d36e29586d04755ace417b659bf6cc437bea36b669ff4b731d89e24"} Dec 01 09:33:46 crc kubenswrapper[4873]: I1201 09:33:46.110593 4873 scope.go:117] "RemoveContainer" containerID="f30c0d46b1aeaa6aedf219761bdb926b5662dc6e6cb3acf1194f97b216eeb0d1" Dec 01 09:33:47 crc kubenswrapper[4873]: I1201 09:33:47.779518 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"67d40a29-6c05-4dfa-9caa-457176a83b49","Type":"ContainerStarted","Data":"da66d3eb73c673b0784211948dd748e6b9734899681d06d627ca732026033663"} Dec 01 09:33:47 crc kubenswrapper[4873]: I1201 09:33:47.780512 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"67d40a29-6c05-4dfa-9caa-457176a83b49","Type":"ContainerStarted","Data":"a1f2e7399882e3cb7e4e82cf88e4dd3bbe56ab56d85ee55cd8a69a307d879ac6"} Dec 01 09:33:47 crc kubenswrapper[4873]: I1201 09:33:47.816977 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=3.754752901 podStartE2EDuration="17.816952036s" podCreationTimestamp="2025-12-01 09:33:30 +0000 UTC" firstStartedPulling="2025-12-01 09:33:32.215445922 +0000 UTC m=+3188.117554451" lastFinishedPulling="2025-12-01 09:33:46.277645047 +0000 UTC m=+3202.179753586" observedRunningTime="2025-12-01 09:33:47.807211112 +0000 UTC m=+3203.709319651" watchObservedRunningTime="2025-12-01 09:33:47.816952036 +0000 UTC m=+3203.719060575" Dec 01 09:33:50 crc kubenswrapper[4873]: I1201 09:33:50.348215 4873 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7c885c5cb8-46fsf" podUID="24452fe2-b990-4537-b619-9db436ed5acb" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.242:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.242:8443: connect: connection refused" Dec 01 09:33:50 crc kubenswrapper[4873]: I1201 09:33:50.865269 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Dec 01 09:33:51 crc kubenswrapper[4873]: I1201 09:33:51.761917 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 09:33:51 crc kubenswrapper[4873]: I1201 09:33:51.835921 4873 generic.go:334] "Generic (PLEG): container finished" podID="c7cad829-b344-4e49-b0d4-264750c0ae0f" containerID="515feff768ee635a4bc61549e094ad2a0d6362338a61af5ceeedf238ceb452c4" exitCode=0 Dec 01 09:33:51 crc kubenswrapper[4873]: I1201 09:33:51.835991 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7cad829-b344-4e49-b0d4-264750c0ae0f","Type":"ContainerDied","Data":"515feff768ee635a4bc61549e094ad2a0d6362338a61af5ceeedf238ceb452c4"} Dec 01 09:33:51 crc kubenswrapper[4873]: I1201 09:33:51.836167 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7cad829-b344-4e49-b0d4-264750c0ae0f","Type":"ContainerDied","Data":"9d6c935168502ee2af9ff782bac01b6f4fa38ed351b738b792e8f36c1190b1b3"} Dec 01 09:33:51 crc kubenswrapper[4873]: I1201 09:33:51.836198 4873 scope.go:117] "RemoveContainer" containerID="2bd108f1fad71742ecc62b281d05995aefb751321590b194fa66df222271b3eb" Dec 01 09:33:51 crc kubenswrapper[4873]: I1201 09:33:51.836585 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 09:33:51 crc kubenswrapper[4873]: I1201 09:33:51.883925 4873 scope.go:117] "RemoveContainer" containerID="c91515ebd87bb1e28b9586fe19335671f8d956b8ecdad460712529b7d0846130" Dec 01 09:33:51 crc kubenswrapper[4873]: I1201 09:33:51.909179 4873 scope.go:117] "RemoveContainer" containerID="515feff768ee635a4bc61549e094ad2a0d6362338a61af5ceeedf238ceb452c4" Dec 01 09:33:51 crc kubenswrapper[4873]: I1201 09:33:51.920629 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7cad829-b344-4e49-b0d4-264750c0ae0f-config-data\") pod \"c7cad829-b344-4e49-b0d4-264750c0ae0f\" (UID: \"c7cad829-b344-4e49-b0d4-264750c0ae0f\") " Dec 01 09:33:51 crc kubenswrapper[4873]: I1201 09:33:51.920714 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7cad829-b344-4e49-b0d4-264750c0ae0f-scripts\") pod \"c7cad829-b344-4e49-b0d4-264750c0ae0f\" (UID: \"c7cad829-b344-4e49-b0d4-264750c0ae0f\") " Dec 01 09:33:51 crc kubenswrapper[4873]: I1201 09:33:51.920749 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7cad829-b344-4e49-b0d4-264750c0ae0f-log-httpd\") pod \"c7cad829-b344-4e49-b0d4-264750c0ae0f\" (UID: \"c7cad829-b344-4e49-b0d4-264750c0ae0f\") " Dec 01 09:33:51 crc kubenswrapper[4873]: I1201 09:33:51.920772 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7cad829-b344-4e49-b0d4-264750c0ae0f-ceilometer-tls-certs\") pod \"c7cad829-b344-4e49-b0d4-264750c0ae0f\" (UID: \"c7cad829-b344-4e49-b0d4-264750c0ae0f\") " Dec 01 09:33:51 crc kubenswrapper[4873]: I1201 09:33:51.920795 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c7cad829-b344-4e49-b0d4-264750c0ae0f-sg-core-conf-yaml\") pod \"c7cad829-b344-4e49-b0d4-264750c0ae0f\" (UID: \"c7cad829-b344-4e49-b0d4-264750c0ae0f\") " Dec 01 09:33:51 crc kubenswrapper[4873]: I1201 09:33:51.920997 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7cad829-b344-4e49-b0d4-264750c0ae0f-run-httpd\") pod \"c7cad829-b344-4e49-b0d4-264750c0ae0f\" (UID: \"c7cad829-b344-4e49-b0d4-264750c0ae0f\") " Dec 01 09:33:51 crc kubenswrapper[4873]: I1201 09:33:51.921479 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7srnd\" (UniqueName: \"kubernetes.io/projected/c7cad829-b344-4e49-b0d4-264750c0ae0f-kube-api-access-7srnd\") pod \"c7cad829-b344-4e49-b0d4-264750c0ae0f\" (UID: \"c7cad829-b344-4e49-b0d4-264750c0ae0f\") " Dec 01 09:33:51 crc kubenswrapper[4873]: I1201 09:33:51.921533 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7cad829-b344-4e49-b0d4-264750c0ae0f-combined-ca-bundle\") pod \"c7cad829-b344-4e49-b0d4-264750c0ae0f\" (UID: \"c7cad829-b344-4e49-b0d4-264750c0ae0f\") " Dec 01 09:33:51 crc kubenswrapper[4873]: I1201 09:33:51.922050 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7cad829-b344-4e49-b0d4-264750c0ae0f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c7cad829-b344-4e49-b0d4-264750c0ae0f" (UID: "c7cad829-b344-4e49-b0d4-264750c0ae0f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:33:51 crc kubenswrapper[4873]: I1201 09:33:51.922391 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7cad829-b344-4e49-b0d4-264750c0ae0f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c7cad829-b344-4e49-b0d4-264750c0ae0f" (UID: "c7cad829-b344-4e49-b0d4-264750c0ae0f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:33:51 crc kubenswrapper[4873]: I1201 09:33:51.922924 4873 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7cad829-b344-4e49-b0d4-264750c0ae0f-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 09:33:51 crc kubenswrapper[4873]: I1201 09:33:51.922955 4873 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7cad829-b344-4e49-b0d4-264750c0ae0f-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 09:33:51 crc kubenswrapper[4873]: I1201 09:33:51.929684 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7cad829-b344-4e49-b0d4-264750c0ae0f-kube-api-access-7srnd" (OuterVolumeSpecName: "kube-api-access-7srnd") pod "c7cad829-b344-4e49-b0d4-264750c0ae0f" (UID: "c7cad829-b344-4e49-b0d4-264750c0ae0f"). InnerVolumeSpecName "kube-api-access-7srnd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:33:51 crc kubenswrapper[4873]: I1201 09:33:51.943934 4873 scope.go:117] "RemoveContainer" containerID="f168c0536d36e29586d04755ace417b659bf6cc437bea36b669ff4b731d89e24" Dec 01 09:33:51 crc kubenswrapper[4873]: I1201 09:33:51.944228 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7cad829-b344-4e49-b0d4-264750c0ae0f-scripts" (OuterVolumeSpecName: "scripts") pod "c7cad829-b344-4e49-b0d4-264750c0ae0f" (UID: "c7cad829-b344-4e49-b0d4-264750c0ae0f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:33:51 crc kubenswrapper[4873]: I1201 09:33:51.959742 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7cad829-b344-4e49-b0d4-264750c0ae0f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c7cad829-b344-4e49-b0d4-264750c0ae0f" (UID: "c7cad829-b344-4e49-b0d4-264750c0ae0f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:33:51 crc kubenswrapper[4873]: I1201 09:33:51.990886 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7cad829-b344-4e49-b0d4-264750c0ae0f-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "c7cad829-b344-4e49-b0d4-264750c0ae0f" (UID: "c7cad829-b344-4e49-b0d4-264750c0ae0f"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.026524 4873 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7cad829-b344-4e49-b0d4-264750c0ae0f-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.026570 4873 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7cad829-b344-4e49-b0d4-264750c0ae0f-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.026587 4873 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c7cad829-b344-4e49-b0d4-264750c0ae0f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.026600 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7srnd\" (UniqueName: \"kubernetes.io/projected/c7cad829-b344-4e49-b0d4-264750c0ae0f-kube-api-access-7srnd\") on node \"crc\" DevicePath \"\"" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.073839 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7cad829-b344-4e49-b0d4-264750c0ae0f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c7cad829-b344-4e49-b0d4-264750c0ae0f" (UID: "c7cad829-b344-4e49-b0d4-264750c0ae0f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.086750 4873 scope.go:117] "RemoveContainer" containerID="2bd108f1fad71742ecc62b281d05995aefb751321590b194fa66df222271b3eb" Dec 01 09:33:52 crc kubenswrapper[4873]: E1201 09:33:52.096308 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2bd108f1fad71742ecc62b281d05995aefb751321590b194fa66df222271b3eb\": container with ID starting with 2bd108f1fad71742ecc62b281d05995aefb751321590b194fa66df222271b3eb not found: ID does not exist" containerID="2bd108f1fad71742ecc62b281d05995aefb751321590b194fa66df222271b3eb" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.096690 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2bd108f1fad71742ecc62b281d05995aefb751321590b194fa66df222271b3eb"} err="failed to get container status \"2bd108f1fad71742ecc62b281d05995aefb751321590b194fa66df222271b3eb\": rpc error: code = NotFound desc = could not find container \"2bd108f1fad71742ecc62b281d05995aefb751321590b194fa66df222271b3eb\": container with ID starting with 2bd108f1fad71742ecc62b281d05995aefb751321590b194fa66df222271b3eb not found: ID does not exist" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.096859 4873 scope.go:117] "RemoveContainer" containerID="c91515ebd87bb1e28b9586fe19335671f8d956b8ecdad460712529b7d0846130" Dec 01 09:33:52 crc kubenswrapper[4873]: E1201 09:33:52.098233 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c91515ebd87bb1e28b9586fe19335671f8d956b8ecdad460712529b7d0846130\": container with ID starting with c91515ebd87bb1e28b9586fe19335671f8d956b8ecdad460712529b7d0846130 not found: ID does not exist" containerID="c91515ebd87bb1e28b9586fe19335671f8d956b8ecdad460712529b7d0846130" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.098311 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c91515ebd87bb1e28b9586fe19335671f8d956b8ecdad460712529b7d0846130"} err="failed to get container status \"c91515ebd87bb1e28b9586fe19335671f8d956b8ecdad460712529b7d0846130\": rpc error: code = NotFound desc = could not find container \"c91515ebd87bb1e28b9586fe19335671f8d956b8ecdad460712529b7d0846130\": container with ID starting with c91515ebd87bb1e28b9586fe19335671f8d956b8ecdad460712529b7d0846130 not found: ID does not exist" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.098354 4873 scope.go:117] "RemoveContainer" containerID="515feff768ee635a4bc61549e094ad2a0d6362338a61af5ceeedf238ceb452c4" Dec 01 09:33:52 crc kubenswrapper[4873]: E1201 09:33:52.098829 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"515feff768ee635a4bc61549e094ad2a0d6362338a61af5ceeedf238ceb452c4\": container with ID starting with 515feff768ee635a4bc61549e094ad2a0d6362338a61af5ceeedf238ceb452c4 not found: ID does not exist" containerID="515feff768ee635a4bc61549e094ad2a0d6362338a61af5ceeedf238ceb452c4" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.098863 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"515feff768ee635a4bc61549e094ad2a0d6362338a61af5ceeedf238ceb452c4"} err="failed to get container status \"515feff768ee635a4bc61549e094ad2a0d6362338a61af5ceeedf238ceb452c4\": rpc error: code = NotFound desc = could not find container \"515feff768ee635a4bc61549e094ad2a0d6362338a61af5ceeedf238ceb452c4\": container with ID starting with 515feff768ee635a4bc61549e094ad2a0d6362338a61af5ceeedf238ceb452c4 not found: ID does not exist" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.098885 4873 scope.go:117] "RemoveContainer" containerID="f168c0536d36e29586d04755ace417b659bf6cc437bea36b669ff4b731d89e24" Dec 01 09:33:52 crc kubenswrapper[4873]: E1201 09:33:52.099191 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f168c0536d36e29586d04755ace417b659bf6cc437bea36b669ff4b731d89e24\": container with ID starting with f168c0536d36e29586d04755ace417b659bf6cc437bea36b669ff4b731d89e24 not found: ID does not exist" containerID="f168c0536d36e29586d04755ace417b659bf6cc437bea36b669ff4b731d89e24" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.099212 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f168c0536d36e29586d04755ace417b659bf6cc437bea36b669ff4b731d89e24"} err="failed to get container status \"f168c0536d36e29586d04755ace417b659bf6cc437bea36b669ff4b731d89e24\": rpc error: code = NotFound desc = could not find container \"f168c0536d36e29586d04755ace417b659bf6cc437bea36b669ff4b731d89e24\": container with ID starting with f168c0536d36e29586d04755ace417b659bf6cc437bea36b669ff4b731d89e24 not found: ID does not exist" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.101836 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7cad829-b344-4e49-b0d4-264750c0ae0f-config-data" (OuterVolumeSpecName: "config-data") pod "c7cad829-b344-4e49-b0d4-264750c0ae0f" (UID: "c7cad829-b344-4e49-b0d4-264750c0ae0f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.129355 4873 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7cad829-b344-4e49-b0d4-264750c0ae0f-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.129416 4873 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7cad829-b344-4e49-b0d4-264750c0ae0f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.179751 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.194964 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.212598 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 09:33:52 crc kubenswrapper[4873]: E1201 09:33:52.213172 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d7e3564-5684-4792-b3b0-57001892f591" containerName="dnsmasq-dns" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.213195 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d7e3564-5684-4792-b3b0-57001892f591" containerName="dnsmasq-dns" Dec 01 09:33:52 crc kubenswrapper[4873]: E1201 09:33:52.213221 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7cad829-b344-4e49-b0d4-264750c0ae0f" containerName="ceilometer-notification-agent" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.213229 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7cad829-b344-4e49-b0d4-264750c0ae0f" containerName="ceilometer-notification-agent" Dec 01 09:33:52 crc kubenswrapper[4873]: E1201 09:33:52.213237 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d7e3564-5684-4792-b3b0-57001892f591" containerName="init" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.213243 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d7e3564-5684-4792-b3b0-57001892f591" containerName="init" Dec 01 09:33:52 crc kubenswrapper[4873]: E1201 09:33:52.213271 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7cad829-b344-4e49-b0d4-264750c0ae0f" containerName="sg-core" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.213278 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7cad829-b344-4e49-b0d4-264750c0ae0f" containerName="sg-core" Dec 01 09:33:52 crc kubenswrapper[4873]: E1201 09:33:52.213290 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7cad829-b344-4e49-b0d4-264750c0ae0f" containerName="proxy-httpd" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.213295 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7cad829-b344-4e49-b0d4-264750c0ae0f" containerName="proxy-httpd" Dec 01 09:33:52 crc kubenswrapper[4873]: E1201 09:33:52.213309 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7cad829-b344-4e49-b0d4-264750c0ae0f" containerName="ceilometer-central-agent" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.213314 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7cad829-b344-4e49-b0d4-264750c0ae0f" containerName="ceilometer-central-agent" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.213513 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7cad829-b344-4e49-b0d4-264750c0ae0f" containerName="proxy-httpd" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.213525 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7cad829-b344-4e49-b0d4-264750c0ae0f" containerName="sg-core" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.213544 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7cad829-b344-4e49-b0d4-264750c0ae0f" containerName="ceilometer-central-agent" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.213561 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7cad829-b344-4e49-b0d4-264750c0ae0f" containerName="ceilometer-notification-agent" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.213578 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d7e3564-5684-4792-b3b0-57001892f591" containerName="dnsmasq-dns" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.218971 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.221673 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.228965 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.229086 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.229418 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.333261 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6ccdf302-89b3-4dc2-94c4-45ea2c2ea241-scripts\") pod \"ceilometer-0\" (UID: \"6ccdf302-89b3-4dc2-94c4-45ea2c2ea241\") " pod="openstack/ceilometer-0" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.333316 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ccdf302-89b3-4dc2-94c4-45ea2c2ea241-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6ccdf302-89b3-4dc2-94c4-45ea2c2ea241\") " pod="openstack/ceilometer-0" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.333341 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ccdf302-89b3-4dc2-94c4-45ea2c2ea241-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"6ccdf302-89b3-4dc2-94c4-45ea2c2ea241\") " pod="openstack/ceilometer-0" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.333368 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ccdf302-89b3-4dc2-94c4-45ea2c2ea241-config-data\") pod \"ceilometer-0\" (UID: \"6ccdf302-89b3-4dc2-94c4-45ea2c2ea241\") " pod="openstack/ceilometer-0" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.333410 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6ccdf302-89b3-4dc2-94c4-45ea2c2ea241-run-httpd\") pod \"ceilometer-0\" (UID: \"6ccdf302-89b3-4dc2-94c4-45ea2c2ea241\") " pod="openstack/ceilometer-0" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.333428 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdjnt\" (UniqueName: \"kubernetes.io/projected/6ccdf302-89b3-4dc2-94c4-45ea2c2ea241-kube-api-access-rdjnt\") pod \"ceilometer-0\" (UID: \"6ccdf302-89b3-4dc2-94c4-45ea2c2ea241\") " pod="openstack/ceilometer-0" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.333477 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6ccdf302-89b3-4dc2-94c4-45ea2c2ea241-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6ccdf302-89b3-4dc2-94c4-45ea2c2ea241\") " pod="openstack/ceilometer-0" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.333565 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6ccdf302-89b3-4dc2-94c4-45ea2c2ea241-log-httpd\") pod \"ceilometer-0\" (UID: \"6ccdf302-89b3-4dc2-94c4-45ea2c2ea241\") " pod="openstack/ceilometer-0" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.439567 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6ccdf302-89b3-4dc2-94c4-45ea2c2ea241-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6ccdf302-89b3-4dc2-94c4-45ea2c2ea241\") " pod="openstack/ceilometer-0" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.439774 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6ccdf302-89b3-4dc2-94c4-45ea2c2ea241-log-httpd\") pod \"ceilometer-0\" (UID: \"6ccdf302-89b3-4dc2-94c4-45ea2c2ea241\") " pod="openstack/ceilometer-0" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.439885 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6ccdf302-89b3-4dc2-94c4-45ea2c2ea241-scripts\") pod \"ceilometer-0\" (UID: \"6ccdf302-89b3-4dc2-94c4-45ea2c2ea241\") " pod="openstack/ceilometer-0" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.439922 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ccdf302-89b3-4dc2-94c4-45ea2c2ea241-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6ccdf302-89b3-4dc2-94c4-45ea2c2ea241\") " pod="openstack/ceilometer-0" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.439955 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ccdf302-89b3-4dc2-94c4-45ea2c2ea241-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"6ccdf302-89b3-4dc2-94c4-45ea2c2ea241\") " pod="openstack/ceilometer-0" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.439988 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ccdf302-89b3-4dc2-94c4-45ea2c2ea241-config-data\") pod \"ceilometer-0\" (UID: \"6ccdf302-89b3-4dc2-94c4-45ea2c2ea241\") " pod="openstack/ceilometer-0" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.440081 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdjnt\" (UniqueName: \"kubernetes.io/projected/6ccdf302-89b3-4dc2-94c4-45ea2c2ea241-kube-api-access-rdjnt\") pod \"ceilometer-0\" (UID: \"6ccdf302-89b3-4dc2-94c4-45ea2c2ea241\") " pod="openstack/ceilometer-0" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.440101 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6ccdf302-89b3-4dc2-94c4-45ea2c2ea241-run-httpd\") pod \"ceilometer-0\" (UID: \"6ccdf302-89b3-4dc2-94c4-45ea2c2ea241\") " pod="openstack/ceilometer-0" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.440469 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6ccdf302-89b3-4dc2-94c4-45ea2c2ea241-log-httpd\") pod \"ceilometer-0\" (UID: \"6ccdf302-89b3-4dc2-94c4-45ea2c2ea241\") " pod="openstack/ceilometer-0" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.440498 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6ccdf302-89b3-4dc2-94c4-45ea2c2ea241-run-httpd\") pod \"ceilometer-0\" (UID: \"6ccdf302-89b3-4dc2-94c4-45ea2c2ea241\") " pod="openstack/ceilometer-0" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.441745 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7cad829-b344-4e49-b0d4-264750c0ae0f" path="/var/lib/kubelet/pods/c7cad829-b344-4e49-b0d4-264750c0ae0f/volumes" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.444756 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6ccdf302-89b3-4dc2-94c4-45ea2c2ea241-scripts\") pod \"ceilometer-0\" (UID: \"6ccdf302-89b3-4dc2-94c4-45ea2c2ea241\") " pod="openstack/ceilometer-0" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.445007 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ccdf302-89b3-4dc2-94c4-45ea2c2ea241-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"6ccdf302-89b3-4dc2-94c4-45ea2c2ea241\") " pod="openstack/ceilometer-0" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.445095 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6ccdf302-89b3-4dc2-94c4-45ea2c2ea241-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6ccdf302-89b3-4dc2-94c4-45ea2c2ea241\") " pod="openstack/ceilometer-0" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.446305 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ccdf302-89b3-4dc2-94c4-45ea2c2ea241-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6ccdf302-89b3-4dc2-94c4-45ea2c2ea241\") " pod="openstack/ceilometer-0" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.448095 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ccdf302-89b3-4dc2-94c4-45ea2c2ea241-config-data\") pod \"ceilometer-0\" (UID: \"6ccdf302-89b3-4dc2-94c4-45ea2c2ea241\") " pod="openstack/ceilometer-0" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.464441 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdjnt\" (UniqueName: \"kubernetes.io/projected/6ccdf302-89b3-4dc2-94c4-45ea2c2ea241-kube-api-access-rdjnt\") pod \"ceilometer-0\" (UID: \"6ccdf302-89b3-4dc2-94c4-45ea2c2ea241\") " pod="openstack/ceilometer-0" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.542810 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.845524 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Dec 01 09:33:52 crc kubenswrapper[4873]: I1201 09:33:52.930638 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Dec 01 09:33:53 crc kubenswrapper[4873]: I1201 09:33:53.083149 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 09:33:53 crc kubenswrapper[4873]: I1201 09:33:53.866379 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="15f823d3-d00f-4c94-bd11-3cd81c911b10" containerName="manila-scheduler" containerID="cri-o://8896e391709e5eb0006fc456e6db831869adf7e5269299dbc0f866dfa489dc2a" gracePeriod=30 Dec 01 09:33:53 crc kubenswrapper[4873]: I1201 09:33:53.866641 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6ccdf302-89b3-4dc2-94c4-45ea2c2ea241","Type":"ContainerStarted","Data":"cdbf3819b111f4a7e7043fc161fd71bb297f8c9a1b45452ae3f0cef5c7c53427"} Dec 01 09:33:53 crc kubenswrapper[4873]: I1201 09:33:53.867041 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="15f823d3-d00f-4c94-bd11-3cd81c911b10" containerName="probe" containerID="cri-o://a1c44d8a9cc1062f2ff31dc42aacc968640cd8481b9aa168ed813adcdd26fc36" gracePeriod=30 Dec 01 09:33:54 crc kubenswrapper[4873]: I1201 09:33:54.881606 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6ccdf302-89b3-4dc2-94c4-45ea2c2ea241","Type":"ContainerStarted","Data":"8646db1a32662f0ca656989a2e8d990eedf11690c690bde9ed9b41709d4c5f23"} Dec 01 09:33:54 crc kubenswrapper[4873]: I1201 09:33:54.884524 4873 generic.go:334] "Generic (PLEG): container finished" podID="15f823d3-d00f-4c94-bd11-3cd81c911b10" containerID="a1c44d8a9cc1062f2ff31dc42aacc968640cd8481b9aa168ed813adcdd26fc36" exitCode=0 Dec 01 09:33:54 crc kubenswrapper[4873]: I1201 09:33:54.884561 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"15f823d3-d00f-4c94-bd11-3cd81c911b10","Type":"ContainerDied","Data":"a1c44d8a9cc1062f2ff31dc42aacc968640cd8481b9aa168ed813adcdd26fc36"} Dec 01 09:33:55 crc kubenswrapper[4873]: I1201 09:33:55.896915 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6ccdf302-89b3-4dc2-94c4-45ea2c2ea241","Type":"ContainerStarted","Data":"8c4b1b7ae36d45021a90aa8cd746abadef2e4d85146d9b80604202df220a9161"} Dec 01 09:33:56 crc kubenswrapper[4873]: I1201 09:33:56.910705 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6ccdf302-89b3-4dc2-94c4-45ea2c2ea241","Type":"ContainerStarted","Data":"b3e32985cc5aa0f7d3a3290becca8646d4c13cad6fbb4515a28844d67120dcdd"} Dec 01 09:33:57 crc kubenswrapper[4873]: I1201 09:33:57.933063 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6ccdf302-89b3-4dc2-94c4-45ea2c2ea241","Type":"ContainerStarted","Data":"608cd58048bfee52faaea1a44281cc32a868c31f2a7dff88113dbd3fcff90fa5"} Dec 01 09:33:57 crc kubenswrapper[4873]: I1201 09:33:57.933725 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 09:33:57 crc kubenswrapper[4873]: I1201 09:33:57.938885 4873 generic.go:334] "Generic (PLEG): container finished" podID="15f823d3-d00f-4c94-bd11-3cd81c911b10" containerID="8896e391709e5eb0006fc456e6db831869adf7e5269299dbc0f866dfa489dc2a" exitCode=0 Dec 01 09:33:57 crc kubenswrapper[4873]: I1201 09:33:57.938944 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"15f823d3-d00f-4c94-bd11-3cd81c911b10","Type":"ContainerDied","Data":"8896e391709e5eb0006fc456e6db831869adf7e5269299dbc0f866dfa489dc2a"} Dec 01 09:33:57 crc kubenswrapper[4873]: I1201 09:33:57.967600 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.625240137 podStartE2EDuration="5.967573228s" podCreationTimestamp="2025-12-01 09:33:52 +0000 UTC" firstStartedPulling="2025-12-01 09:33:53.090240559 +0000 UTC m=+3208.992349108" lastFinishedPulling="2025-12-01 09:33:57.43257366 +0000 UTC m=+3213.334682199" observedRunningTime="2025-12-01 09:33:57.960767288 +0000 UTC m=+3213.862875847" watchObservedRunningTime="2025-12-01 09:33:57.967573228 +0000 UTC m=+3213.869681767" Dec 01 09:33:58 crc kubenswrapper[4873]: I1201 09:33:58.240190 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 01 09:33:58 crc kubenswrapper[4873]: I1201 09:33:58.394962 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rv9gc\" (UniqueName: \"kubernetes.io/projected/15f823d3-d00f-4c94-bd11-3cd81c911b10-kube-api-access-rv9gc\") pod \"15f823d3-d00f-4c94-bd11-3cd81c911b10\" (UID: \"15f823d3-d00f-4c94-bd11-3cd81c911b10\") " Dec 01 09:33:58 crc kubenswrapper[4873]: I1201 09:33:58.395417 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/15f823d3-d00f-4c94-bd11-3cd81c911b10-config-data-custom\") pod \"15f823d3-d00f-4c94-bd11-3cd81c911b10\" (UID: \"15f823d3-d00f-4c94-bd11-3cd81c911b10\") " Dec 01 09:33:58 crc kubenswrapper[4873]: I1201 09:33:58.395487 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15f823d3-d00f-4c94-bd11-3cd81c911b10-combined-ca-bundle\") pod \"15f823d3-d00f-4c94-bd11-3cd81c911b10\" (UID: \"15f823d3-d00f-4c94-bd11-3cd81c911b10\") " Dec 01 09:33:58 crc kubenswrapper[4873]: I1201 09:33:58.395529 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/15f823d3-d00f-4c94-bd11-3cd81c911b10-etc-machine-id\") pod \"15f823d3-d00f-4c94-bd11-3cd81c911b10\" (UID: \"15f823d3-d00f-4c94-bd11-3cd81c911b10\") " Dec 01 09:33:58 crc kubenswrapper[4873]: I1201 09:33:58.395576 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15f823d3-d00f-4c94-bd11-3cd81c911b10-config-data\") pod \"15f823d3-d00f-4c94-bd11-3cd81c911b10\" (UID: \"15f823d3-d00f-4c94-bd11-3cd81c911b10\") " Dec 01 09:33:58 crc kubenswrapper[4873]: I1201 09:33:58.395647 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15f823d3-d00f-4c94-bd11-3cd81c911b10-scripts\") pod \"15f823d3-d00f-4c94-bd11-3cd81c911b10\" (UID: \"15f823d3-d00f-4c94-bd11-3cd81c911b10\") " Dec 01 09:33:58 crc kubenswrapper[4873]: I1201 09:33:58.396818 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/15f823d3-d00f-4c94-bd11-3cd81c911b10-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "15f823d3-d00f-4c94-bd11-3cd81c911b10" (UID: "15f823d3-d00f-4c94-bd11-3cd81c911b10"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:33:58 crc kubenswrapper[4873]: I1201 09:33:58.408169 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15f823d3-d00f-4c94-bd11-3cd81c911b10-scripts" (OuterVolumeSpecName: "scripts") pod "15f823d3-d00f-4c94-bd11-3cd81c911b10" (UID: "15f823d3-d00f-4c94-bd11-3cd81c911b10"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:33:58 crc kubenswrapper[4873]: I1201 09:33:58.410144 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15f823d3-d00f-4c94-bd11-3cd81c911b10-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "15f823d3-d00f-4c94-bd11-3cd81c911b10" (UID: "15f823d3-d00f-4c94-bd11-3cd81c911b10"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:33:58 crc kubenswrapper[4873]: I1201 09:33:58.417459 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15f823d3-d00f-4c94-bd11-3cd81c911b10-kube-api-access-rv9gc" (OuterVolumeSpecName: "kube-api-access-rv9gc") pod "15f823d3-d00f-4c94-bd11-3cd81c911b10" (UID: "15f823d3-d00f-4c94-bd11-3cd81c911b10"). InnerVolumeSpecName "kube-api-access-rv9gc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:33:58 crc kubenswrapper[4873]: I1201 09:33:58.500906 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15f823d3-d00f-4c94-bd11-3cd81c911b10-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "15f823d3-d00f-4c94-bd11-3cd81c911b10" (UID: "15f823d3-d00f-4c94-bd11-3cd81c911b10"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:33:58 crc kubenswrapper[4873]: I1201 09:33:58.501967 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rv9gc\" (UniqueName: \"kubernetes.io/projected/15f823d3-d00f-4c94-bd11-3cd81c911b10-kube-api-access-rv9gc\") on node \"crc\" DevicePath \"\"" Dec 01 09:33:58 crc kubenswrapper[4873]: I1201 09:33:58.501992 4873 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/15f823d3-d00f-4c94-bd11-3cd81c911b10-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 09:33:58 crc kubenswrapper[4873]: I1201 09:33:58.502001 4873 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15f823d3-d00f-4c94-bd11-3cd81c911b10-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:33:58 crc kubenswrapper[4873]: I1201 09:33:58.502023 4873 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/15f823d3-d00f-4c94-bd11-3cd81c911b10-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 01 09:33:58 crc kubenswrapper[4873]: I1201 09:33:58.502035 4873 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15f823d3-d00f-4c94-bd11-3cd81c911b10-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:33:58 crc kubenswrapper[4873]: I1201 09:33:58.574338 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15f823d3-d00f-4c94-bd11-3cd81c911b10-config-data" (OuterVolumeSpecName: "config-data") pod "15f823d3-d00f-4c94-bd11-3cd81c911b10" (UID: "15f823d3-d00f-4c94-bd11-3cd81c911b10"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:33:58 crc kubenswrapper[4873]: I1201 09:33:58.604261 4873 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15f823d3-d00f-4c94-bd11-3cd81c911b10-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:33:58 crc kubenswrapper[4873]: I1201 09:33:58.952340 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"15f823d3-d00f-4c94-bd11-3cd81c911b10","Type":"ContainerDied","Data":"403f125ce7d0ad764382436581a0b6386c4ef4a5b9d96cc03720e541176939c5"} Dec 01 09:33:58 crc kubenswrapper[4873]: I1201 09:33:58.952444 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 01 09:33:58 crc kubenswrapper[4873]: I1201 09:33:58.952466 4873 scope.go:117] "RemoveContainer" containerID="a1c44d8a9cc1062f2ff31dc42aacc968640cd8481b9aa168ed813adcdd26fc36" Dec 01 09:33:58 crc kubenswrapper[4873]: I1201 09:33:58.998217 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Dec 01 09:33:58 crc kubenswrapper[4873]: I1201 09:33:58.999419 4873 scope.go:117] "RemoveContainer" containerID="8896e391709e5eb0006fc456e6db831869adf7e5269299dbc0f866dfa489dc2a" Dec 01 09:33:59 crc kubenswrapper[4873]: I1201 09:33:59.011797 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-scheduler-0"] Dec 01 09:33:59 crc kubenswrapper[4873]: I1201 09:33:59.073218 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Dec 01 09:33:59 crc kubenswrapper[4873]: E1201 09:33:59.074739 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15f823d3-d00f-4c94-bd11-3cd81c911b10" containerName="probe" Dec 01 09:33:59 crc kubenswrapper[4873]: I1201 09:33:59.074769 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="15f823d3-d00f-4c94-bd11-3cd81c911b10" containerName="probe" Dec 01 09:33:59 crc kubenswrapper[4873]: E1201 09:33:59.074849 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15f823d3-d00f-4c94-bd11-3cd81c911b10" containerName="manila-scheduler" Dec 01 09:33:59 crc kubenswrapper[4873]: I1201 09:33:59.074860 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="15f823d3-d00f-4c94-bd11-3cd81c911b10" containerName="manila-scheduler" Dec 01 09:33:59 crc kubenswrapper[4873]: I1201 09:33:59.075409 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="15f823d3-d00f-4c94-bd11-3cd81c911b10" containerName="probe" Dec 01 09:33:59 crc kubenswrapper[4873]: I1201 09:33:59.075457 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="15f823d3-d00f-4c94-bd11-3cd81c911b10" containerName="manila-scheduler" Dec 01 09:33:59 crc kubenswrapper[4873]: I1201 09:33:59.084407 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 01 09:33:59 crc kubenswrapper[4873]: I1201 09:33:59.088203 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Dec 01 09:33:59 crc kubenswrapper[4873]: I1201 09:33:59.122277 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f333cac-7ae8-4563-a71e-3439f5b7717d-config-data\") pod \"manila-scheduler-0\" (UID: \"3f333cac-7ae8-4563-a71e-3439f5b7717d\") " pod="openstack/manila-scheduler-0" Dec 01 09:33:59 crc kubenswrapper[4873]: I1201 09:33:59.122774 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3f333cac-7ae8-4563-a71e-3439f5b7717d-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"3f333cac-7ae8-4563-a71e-3439f5b7717d\") " pod="openstack/manila-scheduler-0" Dec 01 09:33:59 crc kubenswrapper[4873]: I1201 09:33:59.122817 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3f333cac-7ae8-4563-a71e-3439f5b7717d-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"3f333cac-7ae8-4563-a71e-3439f5b7717d\") " pod="openstack/manila-scheduler-0" Dec 01 09:33:59 crc kubenswrapper[4873]: I1201 09:33:59.122990 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5s8gf\" (UniqueName: \"kubernetes.io/projected/3f333cac-7ae8-4563-a71e-3439f5b7717d-kube-api-access-5s8gf\") pod \"manila-scheduler-0\" (UID: \"3f333cac-7ae8-4563-a71e-3439f5b7717d\") " pod="openstack/manila-scheduler-0" Dec 01 09:33:59 crc kubenswrapper[4873]: I1201 09:33:59.123096 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f333cac-7ae8-4563-a71e-3439f5b7717d-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"3f333cac-7ae8-4563-a71e-3439f5b7717d\") " pod="openstack/manila-scheduler-0" Dec 01 09:33:59 crc kubenswrapper[4873]: I1201 09:33:59.123248 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f333cac-7ae8-4563-a71e-3439f5b7717d-scripts\") pod \"manila-scheduler-0\" (UID: \"3f333cac-7ae8-4563-a71e-3439f5b7717d\") " pod="openstack/manila-scheduler-0" Dec 01 09:33:59 crc kubenswrapper[4873]: I1201 09:33:59.127379 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Dec 01 09:33:59 crc kubenswrapper[4873]: I1201 09:33:59.226238 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5s8gf\" (UniqueName: \"kubernetes.io/projected/3f333cac-7ae8-4563-a71e-3439f5b7717d-kube-api-access-5s8gf\") pod \"manila-scheduler-0\" (UID: \"3f333cac-7ae8-4563-a71e-3439f5b7717d\") " pod="openstack/manila-scheduler-0" Dec 01 09:33:59 crc kubenswrapper[4873]: I1201 09:33:59.226360 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f333cac-7ae8-4563-a71e-3439f5b7717d-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"3f333cac-7ae8-4563-a71e-3439f5b7717d\") " pod="openstack/manila-scheduler-0" Dec 01 09:33:59 crc kubenswrapper[4873]: I1201 09:33:59.226456 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f333cac-7ae8-4563-a71e-3439f5b7717d-scripts\") pod \"manila-scheduler-0\" (UID: \"3f333cac-7ae8-4563-a71e-3439f5b7717d\") " pod="openstack/manila-scheduler-0" Dec 01 09:33:59 crc kubenswrapper[4873]: I1201 09:33:59.226512 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f333cac-7ae8-4563-a71e-3439f5b7717d-config-data\") pod \"manila-scheduler-0\" (UID: \"3f333cac-7ae8-4563-a71e-3439f5b7717d\") " pod="openstack/manila-scheduler-0" Dec 01 09:33:59 crc kubenswrapper[4873]: I1201 09:33:59.226556 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3f333cac-7ae8-4563-a71e-3439f5b7717d-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"3f333cac-7ae8-4563-a71e-3439f5b7717d\") " pod="openstack/manila-scheduler-0" Dec 01 09:33:59 crc kubenswrapper[4873]: I1201 09:33:59.226591 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3f333cac-7ae8-4563-a71e-3439f5b7717d-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"3f333cac-7ae8-4563-a71e-3439f5b7717d\") " pod="openstack/manila-scheduler-0" Dec 01 09:33:59 crc kubenswrapper[4873]: I1201 09:33:59.226718 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3f333cac-7ae8-4563-a71e-3439f5b7717d-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"3f333cac-7ae8-4563-a71e-3439f5b7717d\") " pod="openstack/manila-scheduler-0" Dec 01 09:33:59 crc kubenswrapper[4873]: I1201 09:33:59.234734 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f333cac-7ae8-4563-a71e-3439f5b7717d-scripts\") pod \"manila-scheduler-0\" (UID: \"3f333cac-7ae8-4563-a71e-3439f5b7717d\") " pod="openstack/manila-scheduler-0" Dec 01 09:33:59 crc kubenswrapper[4873]: I1201 09:33:59.234928 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f333cac-7ae8-4563-a71e-3439f5b7717d-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"3f333cac-7ae8-4563-a71e-3439f5b7717d\") " pod="openstack/manila-scheduler-0" Dec 01 09:33:59 crc kubenswrapper[4873]: I1201 09:33:59.238940 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f333cac-7ae8-4563-a71e-3439f5b7717d-config-data\") pod \"manila-scheduler-0\" (UID: \"3f333cac-7ae8-4563-a71e-3439f5b7717d\") " pod="openstack/manila-scheduler-0" Dec 01 09:33:59 crc kubenswrapper[4873]: I1201 09:33:59.243837 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3f333cac-7ae8-4563-a71e-3439f5b7717d-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"3f333cac-7ae8-4563-a71e-3439f5b7717d\") " pod="openstack/manila-scheduler-0" Dec 01 09:33:59 crc kubenswrapper[4873]: I1201 09:33:59.246310 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5s8gf\" (UniqueName: \"kubernetes.io/projected/3f333cac-7ae8-4563-a71e-3439f5b7717d-kube-api-access-5s8gf\") pod \"manila-scheduler-0\" (UID: \"3f333cac-7ae8-4563-a71e-3439f5b7717d\") " pod="openstack/manila-scheduler-0" Dec 01 09:33:59 crc kubenswrapper[4873]: I1201 09:33:59.423244 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 01 09:33:59 crc kubenswrapper[4873]: I1201 09:33:59.862305 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/manila-api-0" Dec 01 09:33:59 crc kubenswrapper[4873]: W1201 09:33:59.976266 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3f333cac_7ae8_4563_a71e_3439f5b7717d.slice/crio-95bef17010262b91eca66196c06a925a10ac973760f9f28d62bdb4aa27878d8e WatchSource:0}: Error finding container 95bef17010262b91eca66196c06a925a10ac973760f9f28d62bdb4aa27878d8e: Status 404 returned error can't find the container with id 95bef17010262b91eca66196c06a925a10ac973760f9f28d62bdb4aa27878d8e Dec 01 09:33:59 crc kubenswrapper[4873]: I1201 09:33:59.982786 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Dec 01 09:34:00 crc kubenswrapper[4873]: I1201 09:34:00.347640 4873 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7c885c5cb8-46fsf" podUID="24452fe2-b990-4537-b619-9db436ed5acb" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.242:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.242:8443: connect: connection refused" Dec 01 09:34:00 crc kubenswrapper[4873]: I1201 09:34:00.447226 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15f823d3-d00f-4c94-bd11-3cd81c911b10" path="/var/lib/kubelet/pods/15f823d3-d00f-4c94-bd11-3cd81c911b10/volumes" Dec 01 09:34:01 crc kubenswrapper[4873]: I1201 09:34:01.039385 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"3f333cac-7ae8-4563-a71e-3439f5b7717d","Type":"ContainerStarted","Data":"dd4928b9e64b87cc2c1e1d60e896ee40c5a3610273d5367cdaa92a563b4df9d0"} Dec 01 09:34:01 crc kubenswrapper[4873]: I1201 09:34:01.039975 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"3f333cac-7ae8-4563-a71e-3439f5b7717d","Type":"ContainerStarted","Data":"95bef17010262b91eca66196c06a925a10ac973760f9f28d62bdb4aa27878d8e"} Dec 01 09:34:02 crc kubenswrapper[4873]: I1201 09:34:02.054262 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"3f333cac-7ae8-4563-a71e-3439f5b7717d","Type":"ContainerStarted","Data":"f82cc5a5c4d08aae09aa3ffc331caf41e7aacb804f79f43fcd151fac01aa4642"} Dec 01 09:34:02 crc kubenswrapper[4873]: I1201 09:34:02.088146 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=3.08812205 podStartE2EDuration="3.08812205s" podCreationTimestamp="2025-12-01 09:33:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:34:02.082767596 +0000 UTC m=+3217.984876145" watchObservedRunningTime="2025-12-01 09:34:02.08812205 +0000 UTC m=+3217.990230579" Dec 01 09:34:02 crc kubenswrapper[4873]: I1201 09:34:02.705172 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Dec 01 09:34:02 crc kubenswrapper[4873]: I1201 09:34:02.782930 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Dec 01 09:34:03 crc kubenswrapper[4873]: I1201 09:34:03.069291 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="67d40a29-6c05-4dfa-9caa-457176a83b49" containerName="manila-share" containerID="cri-o://a1f2e7399882e3cb7e4e82cf88e4dd3bbe56ab56d85ee55cd8a69a307d879ac6" gracePeriod=30 Dec 01 09:34:03 crc kubenswrapper[4873]: I1201 09:34:03.069380 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="67d40a29-6c05-4dfa-9caa-457176a83b49" containerName="probe" containerID="cri-o://da66d3eb73c673b0784211948dd748e6b9734899681d06d627ca732026033663" gracePeriod=30 Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.059680 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.120683 4873 generic.go:334] "Generic (PLEG): container finished" podID="67d40a29-6c05-4dfa-9caa-457176a83b49" containerID="da66d3eb73c673b0784211948dd748e6b9734899681d06d627ca732026033663" exitCode=0 Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.120745 4873 generic.go:334] "Generic (PLEG): container finished" podID="67d40a29-6c05-4dfa-9caa-457176a83b49" containerID="a1f2e7399882e3cb7e4e82cf88e4dd3bbe56ab56d85ee55cd8a69a307d879ac6" exitCode=1 Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.120773 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"67d40a29-6c05-4dfa-9caa-457176a83b49","Type":"ContainerDied","Data":"da66d3eb73c673b0784211948dd748e6b9734899681d06d627ca732026033663"} Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.120824 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"67d40a29-6c05-4dfa-9caa-457176a83b49","Type":"ContainerDied","Data":"a1f2e7399882e3cb7e4e82cf88e4dd3bbe56ab56d85ee55cd8a69a307d879ac6"} Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.120835 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"67d40a29-6c05-4dfa-9caa-457176a83b49","Type":"ContainerDied","Data":"3c37578b4a689e47bba8e7ab299ba0e86eb630ae21dca7997fa71c0547581d8a"} Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.120854 4873 scope.go:117] "RemoveContainer" containerID="da66d3eb73c673b0784211948dd748e6b9734899681d06d627ca732026033663" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.121077 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.167518 4873 scope.go:117] "RemoveContainer" containerID="a1f2e7399882e3cb7e4e82cf88e4dd3bbe56ab56d85ee55cd8a69a307d879ac6" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.173204 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/67d40a29-6c05-4dfa-9caa-457176a83b49-ceph\") pod \"67d40a29-6c05-4dfa-9caa-457176a83b49\" (UID: \"67d40a29-6c05-4dfa-9caa-457176a83b49\") " Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.173334 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67d40a29-6c05-4dfa-9caa-457176a83b49-config-data-custom\") pod \"67d40a29-6c05-4dfa-9caa-457176a83b49\" (UID: \"67d40a29-6c05-4dfa-9caa-457176a83b49\") " Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.173493 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67d40a29-6c05-4dfa-9caa-457176a83b49-config-data\") pod \"67d40a29-6c05-4dfa-9caa-457176a83b49\" (UID: \"67d40a29-6c05-4dfa-9caa-457176a83b49\") " Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.173597 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8h49t\" (UniqueName: \"kubernetes.io/projected/67d40a29-6c05-4dfa-9caa-457176a83b49-kube-api-access-8h49t\") pod \"67d40a29-6c05-4dfa-9caa-457176a83b49\" (UID: \"67d40a29-6c05-4dfa-9caa-457176a83b49\") " Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.173655 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/67d40a29-6c05-4dfa-9caa-457176a83b49-etc-machine-id\") pod \"67d40a29-6c05-4dfa-9caa-457176a83b49\" (UID: \"67d40a29-6c05-4dfa-9caa-457176a83b49\") " Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.173704 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/67d40a29-6c05-4dfa-9caa-457176a83b49-var-lib-manila\") pod \"67d40a29-6c05-4dfa-9caa-457176a83b49\" (UID: \"67d40a29-6c05-4dfa-9caa-457176a83b49\") " Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.173757 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67d40a29-6c05-4dfa-9caa-457176a83b49-scripts\") pod \"67d40a29-6c05-4dfa-9caa-457176a83b49\" (UID: \"67d40a29-6c05-4dfa-9caa-457176a83b49\") " Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.173818 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67d40a29-6c05-4dfa-9caa-457176a83b49-combined-ca-bundle\") pod \"67d40a29-6c05-4dfa-9caa-457176a83b49\" (UID: \"67d40a29-6c05-4dfa-9caa-457176a83b49\") " Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.175135 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/67d40a29-6c05-4dfa-9caa-457176a83b49-var-lib-manila" (OuterVolumeSpecName: "var-lib-manila") pod "67d40a29-6c05-4dfa-9caa-457176a83b49" (UID: "67d40a29-6c05-4dfa-9caa-457176a83b49"). InnerVolumeSpecName "var-lib-manila". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.175823 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/67d40a29-6c05-4dfa-9caa-457176a83b49-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "67d40a29-6c05-4dfa-9caa-457176a83b49" (UID: "67d40a29-6c05-4dfa-9caa-457176a83b49"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.182225 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67d40a29-6c05-4dfa-9caa-457176a83b49-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "67d40a29-6c05-4dfa-9caa-457176a83b49" (UID: "67d40a29-6c05-4dfa-9caa-457176a83b49"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.184190 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67d40a29-6c05-4dfa-9caa-457176a83b49-scripts" (OuterVolumeSpecName: "scripts") pod "67d40a29-6c05-4dfa-9caa-457176a83b49" (UID: "67d40a29-6c05-4dfa-9caa-457176a83b49"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.184257 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67d40a29-6c05-4dfa-9caa-457176a83b49-kube-api-access-8h49t" (OuterVolumeSpecName: "kube-api-access-8h49t") pod "67d40a29-6c05-4dfa-9caa-457176a83b49" (UID: "67d40a29-6c05-4dfa-9caa-457176a83b49"). InnerVolumeSpecName "kube-api-access-8h49t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.197566 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67d40a29-6c05-4dfa-9caa-457176a83b49-ceph" (OuterVolumeSpecName: "ceph") pod "67d40a29-6c05-4dfa-9caa-457176a83b49" (UID: "67d40a29-6c05-4dfa-9caa-457176a83b49"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.251770 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67d40a29-6c05-4dfa-9caa-457176a83b49-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "67d40a29-6c05-4dfa-9caa-457176a83b49" (UID: "67d40a29-6c05-4dfa-9caa-457176a83b49"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.264150 4873 scope.go:117] "RemoveContainer" containerID="da66d3eb73c673b0784211948dd748e6b9734899681d06d627ca732026033663" Dec 01 09:34:04 crc kubenswrapper[4873]: E1201 09:34:04.264662 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da66d3eb73c673b0784211948dd748e6b9734899681d06d627ca732026033663\": container with ID starting with da66d3eb73c673b0784211948dd748e6b9734899681d06d627ca732026033663 not found: ID does not exist" containerID="da66d3eb73c673b0784211948dd748e6b9734899681d06d627ca732026033663" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.264700 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da66d3eb73c673b0784211948dd748e6b9734899681d06d627ca732026033663"} err="failed to get container status \"da66d3eb73c673b0784211948dd748e6b9734899681d06d627ca732026033663\": rpc error: code = NotFound desc = could not find container \"da66d3eb73c673b0784211948dd748e6b9734899681d06d627ca732026033663\": container with ID starting with da66d3eb73c673b0784211948dd748e6b9734899681d06d627ca732026033663 not found: ID does not exist" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.264729 4873 scope.go:117] "RemoveContainer" containerID="a1f2e7399882e3cb7e4e82cf88e4dd3bbe56ab56d85ee55cd8a69a307d879ac6" Dec 01 09:34:04 crc kubenswrapper[4873]: E1201 09:34:04.265077 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1f2e7399882e3cb7e4e82cf88e4dd3bbe56ab56d85ee55cd8a69a307d879ac6\": container with ID starting with a1f2e7399882e3cb7e4e82cf88e4dd3bbe56ab56d85ee55cd8a69a307d879ac6 not found: ID does not exist" containerID="a1f2e7399882e3cb7e4e82cf88e4dd3bbe56ab56d85ee55cd8a69a307d879ac6" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.265101 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1f2e7399882e3cb7e4e82cf88e4dd3bbe56ab56d85ee55cd8a69a307d879ac6"} err="failed to get container status \"a1f2e7399882e3cb7e4e82cf88e4dd3bbe56ab56d85ee55cd8a69a307d879ac6\": rpc error: code = NotFound desc = could not find container \"a1f2e7399882e3cb7e4e82cf88e4dd3bbe56ab56d85ee55cd8a69a307d879ac6\": container with ID starting with a1f2e7399882e3cb7e4e82cf88e4dd3bbe56ab56d85ee55cd8a69a307d879ac6 not found: ID does not exist" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.265118 4873 scope.go:117] "RemoveContainer" containerID="da66d3eb73c673b0784211948dd748e6b9734899681d06d627ca732026033663" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.265429 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da66d3eb73c673b0784211948dd748e6b9734899681d06d627ca732026033663"} err="failed to get container status \"da66d3eb73c673b0784211948dd748e6b9734899681d06d627ca732026033663\": rpc error: code = NotFound desc = could not find container \"da66d3eb73c673b0784211948dd748e6b9734899681d06d627ca732026033663\": container with ID starting with da66d3eb73c673b0784211948dd748e6b9734899681d06d627ca732026033663 not found: ID does not exist" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.265450 4873 scope.go:117] "RemoveContainer" containerID="a1f2e7399882e3cb7e4e82cf88e4dd3bbe56ab56d85ee55cd8a69a307d879ac6" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.265708 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1f2e7399882e3cb7e4e82cf88e4dd3bbe56ab56d85ee55cd8a69a307d879ac6"} err="failed to get container status \"a1f2e7399882e3cb7e4e82cf88e4dd3bbe56ab56d85ee55cd8a69a307d879ac6\": rpc error: code = NotFound desc = could not find container \"a1f2e7399882e3cb7e4e82cf88e4dd3bbe56ab56d85ee55cd8a69a307d879ac6\": container with ID starting with a1f2e7399882e3cb7e4e82cf88e4dd3bbe56ab56d85ee55cd8a69a307d879ac6 not found: ID does not exist" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.279162 4873 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/67d40a29-6c05-4dfa-9caa-457176a83b49-ceph\") on node \"crc\" DevicePath \"\"" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.279205 4873 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67d40a29-6c05-4dfa-9caa-457176a83b49-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.279222 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8h49t\" (UniqueName: \"kubernetes.io/projected/67d40a29-6c05-4dfa-9caa-457176a83b49-kube-api-access-8h49t\") on node \"crc\" DevicePath \"\"" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.279238 4873 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/67d40a29-6c05-4dfa-9caa-457176a83b49-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.279248 4873 reconciler_common.go:293] "Volume detached for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/67d40a29-6c05-4dfa-9caa-457176a83b49-var-lib-manila\") on node \"crc\" DevicePath \"\"" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.279256 4873 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67d40a29-6c05-4dfa-9caa-457176a83b49-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.279264 4873 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67d40a29-6c05-4dfa-9caa-457176a83b49-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.319695 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67d40a29-6c05-4dfa-9caa-457176a83b49-config-data" (OuterVolumeSpecName: "config-data") pod "67d40a29-6c05-4dfa-9caa-457176a83b49" (UID: "67d40a29-6c05-4dfa-9caa-457176a83b49"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.392828 4873 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67d40a29-6c05-4dfa-9caa-457176a83b49-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.486198 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.492260 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-share-share1-0"] Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.512955 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Dec 01 09:34:04 crc kubenswrapper[4873]: E1201 09:34:04.513450 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67d40a29-6c05-4dfa-9caa-457176a83b49" containerName="manila-share" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.513475 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="67d40a29-6c05-4dfa-9caa-457176a83b49" containerName="manila-share" Dec 01 09:34:04 crc kubenswrapper[4873]: E1201 09:34:04.513513 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67d40a29-6c05-4dfa-9caa-457176a83b49" containerName="probe" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.513522 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="67d40a29-6c05-4dfa-9caa-457176a83b49" containerName="probe" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.513765 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="67d40a29-6c05-4dfa-9caa-457176a83b49" containerName="manila-share" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.513789 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="67d40a29-6c05-4dfa-9caa-457176a83b49" containerName="probe" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.515003 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.518474 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.542319 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.596721 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f96788d-8f75-490b-99b7-08096baaab34-scripts\") pod \"manila-share-share1-0\" (UID: \"7f96788d-8f75-490b-99b7-08096baaab34\") " pod="openstack/manila-share-share1-0" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.596776 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7f96788d-8f75-490b-99b7-08096baaab34-ceph\") pod \"manila-share-share1-0\" (UID: \"7f96788d-8f75-490b-99b7-08096baaab34\") " pod="openstack/manila-share-share1-0" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.596918 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f96788d-8f75-490b-99b7-08096baaab34-config-data\") pod \"manila-share-share1-0\" (UID: \"7f96788d-8f75-490b-99b7-08096baaab34\") " pod="openstack/manila-share-share1-0" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.597064 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7f96788d-8f75-490b-99b7-08096baaab34-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"7f96788d-8f75-490b-99b7-08096baaab34\") " pod="openstack/manila-share-share1-0" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.597589 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mszdg\" (UniqueName: \"kubernetes.io/projected/7f96788d-8f75-490b-99b7-08096baaab34-kube-api-access-mszdg\") pod \"manila-share-share1-0\" (UID: \"7f96788d-8f75-490b-99b7-08096baaab34\") " pod="openstack/manila-share-share1-0" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.597748 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f96788d-8f75-490b-99b7-08096baaab34-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"7f96788d-8f75-490b-99b7-08096baaab34\") " pod="openstack/manila-share-share1-0" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.597800 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7f96788d-8f75-490b-99b7-08096baaab34-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"7f96788d-8f75-490b-99b7-08096baaab34\") " pod="openstack/manila-share-share1-0" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.597880 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/7f96788d-8f75-490b-99b7-08096baaab34-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"7f96788d-8f75-490b-99b7-08096baaab34\") " pod="openstack/manila-share-share1-0" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.699575 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f96788d-8f75-490b-99b7-08096baaab34-config-data\") pod \"manila-share-share1-0\" (UID: \"7f96788d-8f75-490b-99b7-08096baaab34\") " pod="openstack/manila-share-share1-0" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.699709 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7f96788d-8f75-490b-99b7-08096baaab34-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"7f96788d-8f75-490b-99b7-08096baaab34\") " pod="openstack/manila-share-share1-0" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.699855 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mszdg\" (UniqueName: \"kubernetes.io/projected/7f96788d-8f75-490b-99b7-08096baaab34-kube-api-access-mszdg\") pod \"manila-share-share1-0\" (UID: \"7f96788d-8f75-490b-99b7-08096baaab34\") " pod="openstack/manila-share-share1-0" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.699896 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f96788d-8f75-490b-99b7-08096baaab34-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"7f96788d-8f75-490b-99b7-08096baaab34\") " pod="openstack/manila-share-share1-0" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.699925 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7f96788d-8f75-490b-99b7-08096baaab34-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"7f96788d-8f75-490b-99b7-08096baaab34\") " pod="openstack/manila-share-share1-0" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.699952 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/7f96788d-8f75-490b-99b7-08096baaab34-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"7f96788d-8f75-490b-99b7-08096baaab34\") " pod="openstack/manila-share-share1-0" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.699978 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7f96788d-8f75-490b-99b7-08096baaab34-ceph\") pod \"manila-share-share1-0\" (UID: \"7f96788d-8f75-490b-99b7-08096baaab34\") " pod="openstack/manila-share-share1-0" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.700003 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f96788d-8f75-490b-99b7-08096baaab34-scripts\") pod \"manila-share-share1-0\" (UID: \"7f96788d-8f75-490b-99b7-08096baaab34\") " pod="openstack/manila-share-share1-0" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.700346 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7f96788d-8f75-490b-99b7-08096baaab34-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"7f96788d-8f75-490b-99b7-08096baaab34\") " pod="openstack/manila-share-share1-0" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.700423 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/7f96788d-8f75-490b-99b7-08096baaab34-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"7f96788d-8f75-490b-99b7-08096baaab34\") " pod="openstack/manila-share-share1-0" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.704038 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7f96788d-8f75-490b-99b7-08096baaab34-ceph\") pod \"manila-share-share1-0\" (UID: \"7f96788d-8f75-490b-99b7-08096baaab34\") " pod="openstack/manila-share-share1-0" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.704647 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f96788d-8f75-490b-99b7-08096baaab34-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"7f96788d-8f75-490b-99b7-08096baaab34\") " pod="openstack/manila-share-share1-0" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.704758 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f96788d-8f75-490b-99b7-08096baaab34-scripts\") pod \"manila-share-share1-0\" (UID: \"7f96788d-8f75-490b-99b7-08096baaab34\") " pod="openstack/manila-share-share1-0" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.711583 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f96788d-8f75-490b-99b7-08096baaab34-config-data\") pod \"manila-share-share1-0\" (UID: \"7f96788d-8f75-490b-99b7-08096baaab34\") " pod="openstack/manila-share-share1-0" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.713608 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7f96788d-8f75-490b-99b7-08096baaab34-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"7f96788d-8f75-490b-99b7-08096baaab34\") " pod="openstack/manila-share-share1-0" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.729841 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mszdg\" (UniqueName: \"kubernetes.io/projected/7f96788d-8f75-490b-99b7-08096baaab34-kube-api-access-mszdg\") pod \"manila-share-share1-0\" (UID: \"7f96788d-8f75-490b-99b7-08096baaab34\") " pod="openstack/manila-share-share1-0" Dec 01 09:34:04 crc kubenswrapper[4873]: I1201 09:34:04.835465 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 01 09:34:06 crc kubenswrapper[4873]: I1201 09:34:05.595535 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Dec 01 09:34:06 crc kubenswrapper[4873]: I1201 09:34:06.159065 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"7f96788d-8f75-490b-99b7-08096baaab34","Type":"ContainerStarted","Data":"18f9e0a9f8ef6c71417c3de2dd322a7f95833dd1955ad7df30499ce87d7ed73e"} Dec 01 09:34:06 crc kubenswrapper[4873]: I1201 09:34:06.159345 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"7f96788d-8f75-490b-99b7-08096baaab34","Type":"ContainerStarted","Data":"406400a189ac8b92aba20354fea4939459fc4ca4f1284705e63e43790f845ed8"} Dec 01 09:34:06 crc kubenswrapper[4873]: I1201 09:34:06.445167 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67d40a29-6c05-4dfa-9caa-457176a83b49" path="/var/lib/kubelet/pods/67d40a29-6c05-4dfa-9caa-457176a83b49/volumes" Dec 01 09:34:07 crc kubenswrapper[4873]: I1201 09:34:07.174637 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"7f96788d-8f75-490b-99b7-08096baaab34","Type":"ContainerStarted","Data":"7c05ead8f32cf805b7755b6a26ad97740b6e481ffccce506c3e9c60c2da1ad1b"} Dec 01 09:34:07 crc kubenswrapper[4873]: I1201 09:34:07.204693 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=3.204664553 podStartE2EDuration="3.204664553s" podCreationTimestamp="2025-12-01 09:34:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:34:07.198653973 +0000 UTC m=+3223.100762532" watchObservedRunningTime="2025-12-01 09:34:07.204664553 +0000 UTC m=+3223.106773112" Dec 01 09:34:07 crc kubenswrapper[4873]: I1201 09:34:07.830231 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c885c5cb8-46fsf" Dec 01 09:34:07 crc kubenswrapper[4873]: I1201 09:34:07.888673 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/24452fe2-b990-4537-b619-9db436ed5acb-logs\") pod \"24452fe2-b990-4537-b619-9db436ed5acb\" (UID: \"24452fe2-b990-4537-b619-9db436ed5acb\") " Dec 01 09:34:07 crc kubenswrapper[4873]: I1201 09:34:07.888784 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/24452fe2-b990-4537-b619-9db436ed5acb-scripts\") pod \"24452fe2-b990-4537-b619-9db436ed5acb\" (UID: \"24452fe2-b990-4537-b619-9db436ed5acb\") " Dec 01 09:34:07 crc kubenswrapper[4873]: I1201 09:34:07.888835 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/24452fe2-b990-4537-b619-9db436ed5acb-horizon-secret-key\") pod \"24452fe2-b990-4537-b619-9db436ed5acb\" (UID: \"24452fe2-b990-4537-b619-9db436ed5acb\") " Dec 01 09:34:07 crc kubenswrapper[4873]: I1201 09:34:07.888920 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24452fe2-b990-4537-b619-9db436ed5acb-combined-ca-bundle\") pod \"24452fe2-b990-4537-b619-9db436ed5acb\" (UID: \"24452fe2-b990-4537-b619-9db436ed5acb\") " Dec 01 09:34:07 crc kubenswrapper[4873]: I1201 09:34:07.888971 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ldm2p\" (UniqueName: \"kubernetes.io/projected/24452fe2-b990-4537-b619-9db436ed5acb-kube-api-access-ldm2p\") pod \"24452fe2-b990-4537-b619-9db436ed5acb\" (UID: \"24452fe2-b990-4537-b619-9db436ed5acb\") " Dec 01 09:34:07 crc kubenswrapper[4873]: I1201 09:34:07.889007 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/24452fe2-b990-4537-b619-9db436ed5acb-config-data\") pod \"24452fe2-b990-4537-b619-9db436ed5acb\" (UID: \"24452fe2-b990-4537-b619-9db436ed5acb\") " Dec 01 09:34:07 crc kubenswrapper[4873]: I1201 09:34:07.889073 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/24452fe2-b990-4537-b619-9db436ed5acb-horizon-tls-certs\") pod \"24452fe2-b990-4537-b619-9db436ed5acb\" (UID: \"24452fe2-b990-4537-b619-9db436ed5acb\") " Dec 01 09:34:07 crc kubenswrapper[4873]: I1201 09:34:07.889221 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24452fe2-b990-4537-b619-9db436ed5acb-logs" (OuterVolumeSpecName: "logs") pod "24452fe2-b990-4537-b619-9db436ed5acb" (UID: "24452fe2-b990-4537-b619-9db436ed5acb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:34:07 crc kubenswrapper[4873]: I1201 09:34:07.889739 4873 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/24452fe2-b990-4537-b619-9db436ed5acb-logs\") on node \"crc\" DevicePath \"\"" Dec 01 09:34:07 crc kubenswrapper[4873]: I1201 09:34:07.898465 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24452fe2-b990-4537-b619-9db436ed5acb-kube-api-access-ldm2p" (OuterVolumeSpecName: "kube-api-access-ldm2p") pod "24452fe2-b990-4537-b619-9db436ed5acb" (UID: "24452fe2-b990-4537-b619-9db436ed5acb"). InnerVolumeSpecName "kube-api-access-ldm2p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:34:07 crc kubenswrapper[4873]: I1201 09:34:07.901091 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24452fe2-b990-4537-b619-9db436ed5acb-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "24452fe2-b990-4537-b619-9db436ed5acb" (UID: "24452fe2-b990-4537-b619-9db436ed5acb"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:34:07 crc kubenswrapper[4873]: I1201 09:34:07.928680 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24452fe2-b990-4537-b619-9db436ed5acb-scripts" (OuterVolumeSpecName: "scripts") pod "24452fe2-b990-4537-b619-9db436ed5acb" (UID: "24452fe2-b990-4537-b619-9db436ed5acb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:34:07 crc kubenswrapper[4873]: I1201 09:34:07.932876 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24452fe2-b990-4537-b619-9db436ed5acb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "24452fe2-b990-4537-b619-9db436ed5acb" (UID: "24452fe2-b990-4537-b619-9db436ed5acb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:34:07 crc kubenswrapper[4873]: I1201 09:34:07.952066 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24452fe2-b990-4537-b619-9db436ed5acb-config-data" (OuterVolumeSpecName: "config-data") pod "24452fe2-b990-4537-b619-9db436ed5acb" (UID: "24452fe2-b990-4537-b619-9db436ed5acb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:34:07 crc kubenswrapper[4873]: I1201 09:34:07.963635 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24452fe2-b990-4537-b619-9db436ed5acb-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "24452fe2-b990-4537-b619-9db436ed5acb" (UID: "24452fe2-b990-4537-b619-9db436ed5acb"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:34:07 crc kubenswrapper[4873]: I1201 09:34:07.991780 4873 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/24452fe2-b990-4537-b619-9db436ed5acb-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:34:07 crc kubenswrapper[4873]: I1201 09:34:07.991824 4873 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/24452fe2-b990-4537-b619-9db436ed5acb-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 01 09:34:07 crc kubenswrapper[4873]: I1201 09:34:07.991838 4873 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24452fe2-b990-4537-b619-9db436ed5acb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:34:07 crc kubenswrapper[4873]: I1201 09:34:07.991847 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ldm2p\" (UniqueName: \"kubernetes.io/projected/24452fe2-b990-4537-b619-9db436ed5acb-kube-api-access-ldm2p\") on node \"crc\" DevicePath \"\"" Dec 01 09:34:07 crc kubenswrapper[4873]: I1201 09:34:07.991860 4873 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/24452fe2-b990-4537-b619-9db436ed5acb-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:34:07 crc kubenswrapper[4873]: I1201 09:34:07.991868 4873 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/24452fe2-b990-4537-b619-9db436ed5acb-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 09:34:08 crc kubenswrapper[4873]: I1201 09:34:08.204245 4873 generic.go:334] "Generic (PLEG): container finished" podID="24452fe2-b990-4537-b619-9db436ed5acb" containerID="3f684c99d8aa2dce24bd09bca0e8fa379cf0d83ddd7621cb5266dddb613a60a8" exitCode=137 Dec 01 09:34:08 crc kubenswrapper[4873]: I1201 09:34:08.204354 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c885c5cb8-46fsf" event={"ID":"24452fe2-b990-4537-b619-9db436ed5acb","Type":"ContainerDied","Data":"3f684c99d8aa2dce24bd09bca0e8fa379cf0d83ddd7621cb5266dddb613a60a8"} Dec 01 09:34:08 crc kubenswrapper[4873]: I1201 09:34:08.204417 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c885c5cb8-46fsf" event={"ID":"24452fe2-b990-4537-b619-9db436ed5acb","Type":"ContainerDied","Data":"2398a1a8f1a9676fc632643ad733b0d8e0a617d5488c48bdcbe599ea953a68d1"} Dec 01 09:34:08 crc kubenswrapper[4873]: I1201 09:34:08.204439 4873 scope.go:117] "RemoveContainer" containerID="19a1f4707bbeeffe7274e84c66714fc06a3e1466289970827846d71126cb57c3" Dec 01 09:34:08 crc kubenswrapper[4873]: I1201 09:34:08.205783 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c885c5cb8-46fsf" Dec 01 09:34:08 crc kubenswrapper[4873]: I1201 09:34:08.252831 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7c885c5cb8-46fsf"] Dec 01 09:34:08 crc kubenswrapper[4873]: I1201 09:34:08.262911 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7c885c5cb8-46fsf"] Dec 01 09:34:08 crc kubenswrapper[4873]: I1201 09:34:08.404758 4873 scope.go:117] "RemoveContainer" containerID="3f684c99d8aa2dce24bd09bca0e8fa379cf0d83ddd7621cb5266dddb613a60a8" Dec 01 09:34:08 crc kubenswrapper[4873]: I1201 09:34:08.435196 4873 scope.go:117] "RemoveContainer" containerID="19a1f4707bbeeffe7274e84c66714fc06a3e1466289970827846d71126cb57c3" Dec 01 09:34:08 crc kubenswrapper[4873]: E1201 09:34:08.436307 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19a1f4707bbeeffe7274e84c66714fc06a3e1466289970827846d71126cb57c3\": container with ID starting with 19a1f4707bbeeffe7274e84c66714fc06a3e1466289970827846d71126cb57c3 not found: ID does not exist" containerID="19a1f4707bbeeffe7274e84c66714fc06a3e1466289970827846d71126cb57c3" Dec 01 09:34:08 crc kubenswrapper[4873]: I1201 09:34:08.436373 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19a1f4707bbeeffe7274e84c66714fc06a3e1466289970827846d71126cb57c3"} err="failed to get container status \"19a1f4707bbeeffe7274e84c66714fc06a3e1466289970827846d71126cb57c3\": rpc error: code = NotFound desc = could not find container \"19a1f4707bbeeffe7274e84c66714fc06a3e1466289970827846d71126cb57c3\": container with ID starting with 19a1f4707bbeeffe7274e84c66714fc06a3e1466289970827846d71126cb57c3 not found: ID does not exist" Dec 01 09:34:08 crc kubenswrapper[4873]: I1201 09:34:08.436409 4873 scope.go:117] "RemoveContainer" containerID="3f684c99d8aa2dce24bd09bca0e8fa379cf0d83ddd7621cb5266dddb613a60a8" Dec 01 09:34:08 crc kubenswrapper[4873]: E1201 09:34:08.436838 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f684c99d8aa2dce24bd09bca0e8fa379cf0d83ddd7621cb5266dddb613a60a8\": container with ID starting with 3f684c99d8aa2dce24bd09bca0e8fa379cf0d83ddd7621cb5266dddb613a60a8 not found: ID does not exist" containerID="3f684c99d8aa2dce24bd09bca0e8fa379cf0d83ddd7621cb5266dddb613a60a8" Dec 01 09:34:08 crc kubenswrapper[4873]: I1201 09:34:08.436870 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f684c99d8aa2dce24bd09bca0e8fa379cf0d83ddd7621cb5266dddb613a60a8"} err="failed to get container status \"3f684c99d8aa2dce24bd09bca0e8fa379cf0d83ddd7621cb5266dddb613a60a8\": rpc error: code = NotFound desc = could not find container \"3f684c99d8aa2dce24bd09bca0e8fa379cf0d83ddd7621cb5266dddb613a60a8\": container with ID starting with 3f684c99d8aa2dce24bd09bca0e8fa379cf0d83ddd7621cb5266dddb613a60a8 not found: ID does not exist" Dec 01 09:34:08 crc kubenswrapper[4873]: I1201 09:34:08.443621 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24452fe2-b990-4537-b619-9db436ed5acb" path="/var/lib/kubelet/pods/24452fe2-b990-4537-b619-9db436ed5acb/volumes" Dec 01 09:34:09 crc kubenswrapper[4873]: I1201 09:34:09.423738 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Dec 01 09:34:14 crc kubenswrapper[4873]: I1201 09:34:14.835833 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Dec 01 09:34:21 crc kubenswrapper[4873]: I1201 09:34:21.311217 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Dec 01 09:34:22 crc kubenswrapper[4873]: I1201 09:34:22.552118 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 01 09:34:26 crc kubenswrapper[4873]: I1201 09:34:26.596521 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Dec 01 09:35:03 crc kubenswrapper[4873]: I1201 09:35:03.688212 4873 scope.go:117] "RemoveContainer" containerID="ef68547a5f0564f4ebd8555c369cfa7f9132506cf6b066a9d6d8e9138f8eef38" Dec 01 09:35:03 crc kubenswrapper[4873]: I1201 09:35:03.711301 4873 scope.go:117] "RemoveContainer" containerID="a3aae70def07ef8ca54a1bcbbbee1691d453bd0e07d9bbc469320b4aa49f127f" Dec 01 09:35:03 crc kubenswrapper[4873]: I1201 09:35:03.736340 4873 scope.go:117] "RemoveContainer" containerID="14fb93271d6a7e20e0456a629c4a78d3c5d064c6950f2dcb87e7e4d229af009b" Dec 01 09:35:17 crc kubenswrapper[4873]: I1201 09:35:17.703469 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Dec 01 09:35:17 crc kubenswrapper[4873]: E1201 09:35:17.704781 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24452fe2-b990-4537-b619-9db436ed5acb" containerName="horizon-log" Dec 01 09:35:17 crc kubenswrapper[4873]: I1201 09:35:17.704812 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="24452fe2-b990-4537-b619-9db436ed5acb" containerName="horizon-log" Dec 01 09:35:17 crc kubenswrapper[4873]: E1201 09:35:17.704830 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24452fe2-b990-4537-b619-9db436ed5acb" containerName="horizon" Dec 01 09:35:17 crc kubenswrapper[4873]: I1201 09:35:17.704840 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="24452fe2-b990-4537-b619-9db436ed5acb" containerName="horizon" Dec 01 09:35:17 crc kubenswrapper[4873]: I1201 09:35:17.705499 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="24452fe2-b990-4537-b619-9db436ed5acb" containerName="horizon" Dec 01 09:35:17 crc kubenswrapper[4873]: I1201 09:35:17.705529 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="24452fe2-b990-4537-b619-9db436ed5acb" containerName="horizon-log" Dec 01 09:35:17 crc kubenswrapper[4873]: I1201 09:35:17.709532 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 01 09:35:17 crc kubenswrapper[4873]: I1201 09:35:17.712677 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 01 09:35:17 crc kubenswrapper[4873]: I1201 09:35:17.713463 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 01 09:35:17 crc kubenswrapper[4873]: I1201 09:35:17.713710 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 01 09:35:17 crc kubenswrapper[4873]: I1201 09:35:17.713898 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-pqxv2" Dec 01 09:35:17 crc kubenswrapper[4873]: I1201 09:35:17.735030 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 01 09:35:17 crc kubenswrapper[4873]: I1201 09:35:17.811339 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-config-data\") pod \"tempest-tests-tempest\" (UID: \"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1\") " pod="openstack/tempest-tests-tempest" Dec 01 09:35:17 crc kubenswrapper[4873]: I1201 09:35:17.811396 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1\") " pod="openstack/tempest-tests-tempest" Dec 01 09:35:17 crc kubenswrapper[4873]: I1201 09:35:17.811436 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1\") " pod="openstack/tempest-tests-tempest" Dec 01 09:35:17 crc kubenswrapper[4873]: I1201 09:35:17.811453 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1\") " pod="openstack/tempest-tests-tempest" Dec 01 09:35:17 crc kubenswrapper[4873]: I1201 09:35:17.811492 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1\") " pod="openstack/tempest-tests-tempest" Dec 01 09:35:17 crc kubenswrapper[4873]: I1201 09:35:17.811528 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1\") " pod="openstack/tempest-tests-tempest" Dec 01 09:35:17 crc kubenswrapper[4873]: I1201 09:35:17.811549 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1\") " pod="openstack/tempest-tests-tempest" Dec 01 09:35:17 crc kubenswrapper[4873]: I1201 09:35:17.811625 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1\") " pod="openstack/tempest-tests-tempest" Dec 01 09:35:17 crc kubenswrapper[4873]: I1201 09:35:17.811642 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-958mm\" (UniqueName: \"kubernetes.io/projected/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-kube-api-access-958mm\") pod \"tempest-tests-tempest\" (UID: \"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1\") " pod="openstack/tempest-tests-tempest" Dec 01 09:35:17 crc kubenswrapper[4873]: I1201 09:35:17.915700 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-config-data\") pod \"tempest-tests-tempest\" (UID: \"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1\") " pod="openstack/tempest-tests-tempest" Dec 01 09:35:17 crc kubenswrapper[4873]: I1201 09:35:17.913942 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-config-data\") pod \"tempest-tests-tempest\" (UID: \"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1\") " pod="openstack/tempest-tests-tempest" Dec 01 09:35:17 crc kubenswrapper[4873]: I1201 09:35:17.919135 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1\") " pod="openstack/tempest-tests-tempest" Dec 01 09:35:17 crc kubenswrapper[4873]: I1201 09:35:17.919419 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1\") " pod="openstack/tempest-tests-tempest" Dec 01 09:35:17 crc kubenswrapper[4873]: I1201 09:35:17.919573 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1\") " pod="openstack/tempest-tests-tempest" Dec 01 09:35:17 crc kubenswrapper[4873]: I1201 09:35:17.920220 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1\") " pod="openstack/tempest-tests-tempest" Dec 01 09:35:17 crc kubenswrapper[4873]: I1201 09:35:17.920170 4873 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/tempest-tests-tempest" Dec 01 09:35:17 crc kubenswrapper[4873]: I1201 09:35:17.921561 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1\") " pod="openstack/tempest-tests-tempest" Dec 01 09:35:17 crc kubenswrapper[4873]: I1201 09:35:17.921729 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1\") " pod="openstack/tempest-tests-tempest" Dec 01 09:35:17 crc kubenswrapper[4873]: I1201 09:35:17.921779 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1\") " pod="openstack/tempest-tests-tempest" Dec 01 09:35:17 crc kubenswrapper[4873]: I1201 09:35:17.922169 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1\") " pod="openstack/tempest-tests-tempest" Dec 01 09:35:17 crc kubenswrapper[4873]: I1201 09:35:17.922203 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-958mm\" (UniqueName: \"kubernetes.io/projected/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-kube-api-access-958mm\") pod \"tempest-tests-tempest\" (UID: \"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1\") " pod="openstack/tempest-tests-tempest" Dec 01 09:35:17 crc kubenswrapper[4873]: I1201 09:35:17.923631 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1\") " pod="openstack/tempest-tests-tempest" Dec 01 09:35:17 crc kubenswrapper[4873]: I1201 09:35:17.924823 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1\") " pod="openstack/tempest-tests-tempest" Dec 01 09:35:17 crc kubenswrapper[4873]: I1201 09:35:17.936406 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1\") " pod="openstack/tempest-tests-tempest" Dec 01 09:35:17 crc kubenswrapper[4873]: I1201 09:35:17.937121 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1\") " pod="openstack/tempest-tests-tempest" Dec 01 09:35:17 crc kubenswrapper[4873]: I1201 09:35:17.940565 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1\") " pod="openstack/tempest-tests-tempest" Dec 01 09:35:17 crc kubenswrapper[4873]: I1201 09:35:17.946457 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-958mm\" (UniqueName: \"kubernetes.io/projected/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-kube-api-access-958mm\") pod \"tempest-tests-tempest\" (UID: \"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1\") " pod="openstack/tempest-tests-tempest" Dec 01 09:35:17 crc kubenswrapper[4873]: I1201 09:35:17.961919 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1\") " pod="openstack/tempest-tests-tempest" Dec 01 09:35:18 crc kubenswrapper[4873]: I1201 09:35:18.035147 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 01 09:35:18 crc kubenswrapper[4873]: I1201 09:35:18.499429 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 01 09:35:18 crc kubenswrapper[4873]: I1201 09:35:18.946126 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1","Type":"ContainerStarted","Data":"59b498de32d7198342495c719c447abce031a7c6509fe1755f6f3c7946822ed9"} Dec 01 09:35:51 crc kubenswrapper[4873]: E1201 09:35:51.628070 4873 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Dec 01 09:35:51 crc kubenswrapper[4873]: E1201 09:35:51.628704 4873 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-958mm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(cd15a962-a485-48ad-b0af-d7a3ec0e4ab1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 09:35:51 crc kubenswrapper[4873]: E1201 09:35:51.629955 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="cd15a962-a485-48ad-b0af-d7a3ec0e4ab1" Dec 01 09:35:52 crc kubenswrapper[4873]: E1201 09:35:52.365394 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="cd15a962-a485-48ad-b0af-d7a3ec0e4ab1" Dec 01 09:36:01 crc kubenswrapper[4873]: I1201 09:36:01.060108 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:36:01 crc kubenswrapper[4873]: I1201 09:36:01.060591 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:36:06 crc kubenswrapper[4873]: I1201 09:36:06.969081 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 01 09:36:09 crc kubenswrapper[4873]: I1201 09:36:09.086589 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1","Type":"ContainerStarted","Data":"098e20bc0ab355e1003b658643d1f17730f91014cf149b5d9058a091734d8dd6"} Dec 01 09:36:31 crc kubenswrapper[4873]: I1201 09:36:31.059009 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:36:31 crc kubenswrapper[4873]: I1201 09:36:31.059930 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:37:01 crc kubenswrapper[4873]: I1201 09:37:01.059427 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:37:01 crc kubenswrapper[4873]: I1201 09:37:01.060000 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:37:01 crc kubenswrapper[4873]: I1201 09:37:01.060067 4873 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" Dec 01 09:37:01 crc kubenswrapper[4873]: I1201 09:37:01.060619 4873 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0b72d3c91d2960798896e9520de55a4bb91b7db511fab82dd6f7e66083a742f2"} pod="openshift-machine-config-operator/machine-config-daemon-scwpp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 09:37:01 crc kubenswrapper[4873]: I1201 09:37:01.060670 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" containerID="cri-o://0b72d3c91d2960798896e9520de55a4bb91b7db511fab82dd6f7e66083a742f2" gracePeriod=600 Dec 01 09:37:01 crc kubenswrapper[4873]: I1201 09:37:01.624855 4873 generic.go:334] "Generic (PLEG): container finished" podID="fef7b114-0e07-402d-a37b-315c36011f4b" containerID="0b72d3c91d2960798896e9520de55a4bb91b7db511fab82dd6f7e66083a742f2" exitCode=0 Dec 01 09:37:01 crc kubenswrapper[4873]: I1201 09:37:01.624930 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" event={"ID":"fef7b114-0e07-402d-a37b-315c36011f4b","Type":"ContainerDied","Data":"0b72d3c91d2960798896e9520de55a4bb91b7db511fab82dd6f7e66083a742f2"} Dec 01 09:37:01 crc kubenswrapper[4873]: I1201 09:37:01.626133 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" event={"ID":"fef7b114-0e07-402d-a37b-315c36011f4b","Type":"ContainerStarted","Data":"c101914cc0d7462afe57653cc6d66f5335bbd707b5b5acec3231122718624d93"} Dec 01 09:37:01 crc kubenswrapper[4873]: I1201 09:37:01.626232 4873 scope.go:117] "RemoveContainer" containerID="1eb6abed9ec75ad80889e13d075ac04aa13db51de45fa2f980efa98223444319" Dec 01 09:37:01 crc kubenswrapper[4873]: I1201 09:37:01.653649 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=57.211217215 podStartE2EDuration="1m45.65362183s" podCreationTimestamp="2025-12-01 09:35:16 +0000 UTC" firstStartedPulling="2025-12-01 09:35:18.523117868 +0000 UTC m=+3294.425226407" lastFinishedPulling="2025-12-01 09:36:06.965522483 +0000 UTC m=+3342.867631022" observedRunningTime="2025-12-01 09:36:09.122254726 +0000 UTC m=+3345.024363295" watchObservedRunningTime="2025-12-01 09:37:01.65362183 +0000 UTC m=+3397.555730399" Dec 01 09:37:54 crc kubenswrapper[4873]: I1201 09:37:54.285221 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9l6jh"] Dec 01 09:37:54 crc kubenswrapper[4873]: I1201 09:37:54.289440 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9l6jh" Dec 01 09:37:54 crc kubenswrapper[4873]: I1201 09:37:54.308955 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9l6jh"] Dec 01 09:37:54 crc kubenswrapper[4873]: I1201 09:37:54.447983 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a29f28fe-2850-45a0-9b3e-eb84ef30fff3-catalog-content\") pod \"community-operators-9l6jh\" (UID: \"a29f28fe-2850-45a0-9b3e-eb84ef30fff3\") " pod="openshift-marketplace/community-operators-9l6jh" Dec 01 09:37:54 crc kubenswrapper[4873]: I1201 09:37:54.448148 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a29f28fe-2850-45a0-9b3e-eb84ef30fff3-utilities\") pod \"community-operators-9l6jh\" (UID: \"a29f28fe-2850-45a0-9b3e-eb84ef30fff3\") " pod="openshift-marketplace/community-operators-9l6jh" Dec 01 09:37:54 crc kubenswrapper[4873]: I1201 09:37:54.448210 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whhgs\" (UniqueName: \"kubernetes.io/projected/a29f28fe-2850-45a0-9b3e-eb84ef30fff3-kube-api-access-whhgs\") pod \"community-operators-9l6jh\" (UID: \"a29f28fe-2850-45a0-9b3e-eb84ef30fff3\") " pod="openshift-marketplace/community-operators-9l6jh" Dec 01 09:37:54 crc kubenswrapper[4873]: I1201 09:37:54.551655 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a29f28fe-2850-45a0-9b3e-eb84ef30fff3-catalog-content\") pod \"community-operators-9l6jh\" (UID: \"a29f28fe-2850-45a0-9b3e-eb84ef30fff3\") " pod="openshift-marketplace/community-operators-9l6jh" Dec 01 09:37:54 crc kubenswrapper[4873]: I1201 09:37:54.552234 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a29f28fe-2850-45a0-9b3e-eb84ef30fff3-catalog-content\") pod \"community-operators-9l6jh\" (UID: \"a29f28fe-2850-45a0-9b3e-eb84ef30fff3\") " pod="openshift-marketplace/community-operators-9l6jh" Dec 01 09:37:54 crc kubenswrapper[4873]: I1201 09:37:54.552731 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a29f28fe-2850-45a0-9b3e-eb84ef30fff3-utilities\") pod \"community-operators-9l6jh\" (UID: \"a29f28fe-2850-45a0-9b3e-eb84ef30fff3\") " pod="openshift-marketplace/community-operators-9l6jh" Dec 01 09:37:54 crc kubenswrapper[4873]: I1201 09:37:54.553000 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a29f28fe-2850-45a0-9b3e-eb84ef30fff3-utilities\") pod \"community-operators-9l6jh\" (UID: \"a29f28fe-2850-45a0-9b3e-eb84ef30fff3\") " pod="openshift-marketplace/community-operators-9l6jh" Dec 01 09:37:54 crc kubenswrapper[4873]: I1201 09:37:54.553488 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whhgs\" (UniqueName: \"kubernetes.io/projected/a29f28fe-2850-45a0-9b3e-eb84ef30fff3-kube-api-access-whhgs\") pod \"community-operators-9l6jh\" (UID: \"a29f28fe-2850-45a0-9b3e-eb84ef30fff3\") " pod="openshift-marketplace/community-operators-9l6jh" Dec 01 09:37:54 crc kubenswrapper[4873]: I1201 09:37:54.572899 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whhgs\" (UniqueName: \"kubernetes.io/projected/a29f28fe-2850-45a0-9b3e-eb84ef30fff3-kube-api-access-whhgs\") pod \"community-operators-9l6jh\" (UID: \"a29f28fe-2850-45a0-9b3e-eb84ef30fff3\") " pod="openshift-marketplace/community-operators-9l6jh" Dec 01 09:37:54 crc kubenswrapper[4873]: I1201 09:37:54.620627 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9l6jh" Dec 01 09:37:55 crc kubenswrapper[4873]: I1201 09:37:55.172148 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9l6jh"] Dec 01 09:37:56 crc kubenswrapper[4873]: I1201 09:37:56.184892 4873 generic.go:334] "Generic (PLEG): container finished" podID="a29f28fe-2850-45a0-9b3e-eb84ef30fff3" containerID="eb59298f8d036d67723496770ab1fc3b339397f4240423806d6bc2122e61bcbb" exitCode=0 Dec 01 09:37:56 crc kubenswrapper[4873]: I1201 09:37:56.184983 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9l6jh" event={"ID":"a29f28fe-2850-45a0-9b3e-eb84ef30fff3","Type":"ContainerDied","Data":"eb59298f8d036d67723496770ab1fc3b339397f4240423806d6bc2122e61bcbb"} Dec 01 09:37:56 crc kubenswrapper[4873]: I1201 09:37:56.185305 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9l6jh" event={"ID":"a29f28fe-2850-45a0-9b3e-eb84ef30fff3","Type":"ContainerStarted","Data":"b9c4478a0bc193da6c44d6d9a2830a968f72dcf0a11f2468ea6bb18563f51310"} Dec 01 09:37:56 crc kubenswrapper[4873]: I1201 09:37:56.188997 4873 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 09:37:57 crc kubenswrapper[4873]: I1201 09:37:57.200334 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9l6jh" event={"ID":"a29f28fe-2850-45a0-9b3e-eb84ef30fff3","Type":"ContainerStarted","Data":"8d0416829d08beb6b01c389e53dd108adb39754cc7685dfc62c9d561242d0235"} Dec 01 09:37:58 crc kubenswrapper[4873]: I1201 09:37:58.214731 4873 generic.go:334] "Generic (PLEG): container finished" podID="a29f28fe-2850-45a0-9b3e-eb84ef30fff3" containerID="8d0416829d08beb6b01c389e53dd108adb39754cc7685dfc62c9d561242d0235" exitCode=0 Dec 01 09:37:58 crc kubenswrapper[4873]: I1201 09:37:58.214801 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9l6jh" event={"ID":"a29f28fe-2850-45a0-9b3e-eb84ef30fff3","Type":"ContainerDied","Data":"8d0416829d08beb6b01c389e53dd108adb39754cc7685dfc62c9d561242d0235"} Dec 01 09:37:59 crc kubenswrapper[4873]: I1201 09:37:59.227833 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9l6jh" event={"ID":"a29f28fe-2850-45a0-9b3e-eb84ef30fff3","Type":"ContainerStarted","Data":"4bf7bf36e3d0130cfdbc28caf002c5658e1ed940b10d8e2e726abeb5084791a2"} Dec 01 09:37:59 crc kubenswrapper[4873]: I1201 09:37:59.252409 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9l6jh" podStartSLOduration=2.763567741 podStartE2EDuration="5.252381718s" podCreationTimestamp="2025-12-01 09:37:54 +0000 UTC" firstStartedPulling="2025-12-01 09:37:56.188674062 +0000 UTC m=+3452.090782601" lastFinishedPulling="2025-12-01 09:37:58.677488049 +0000 UTC m=+3454.579596578" observedRunningTime="2025-12-01 09:37:59.24490655 +0000 UTC m=+3455.147015099" watchObservedRunningTime="2025-12-01 09:37:59.252381718 +0000 UTC m=+3455.154490267" Dec 01 09:38:04 crc kubenswrapper[4873]: I1201 09:38:04.621442 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9l6jh" Dec 01 09:38:04 crc kubenswrapper[4873]: I1201 09:38:04.622705 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9l6jh" Dec 01 09:38:04 crc kubenswrapper[4873]: I1201 09:38:04.685927 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9l6jh" Dec 01 09:38:05 crc kubenswrapper[4873]: I1201 09:38:05.353707 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9l6jh" Dec 01 09:38:05 crc kubenswrapper[4873]: I1201 09:38:05.432665 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9l6jh"] Dec 01 09:38:07 crc kubenswrapper[4873]: I1201 09:38:07.305092 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9l6jh" podUID="a29f28fe-2850-45a0-9b3e-eb84ef30fff3" containerName="registry-server" containerID="cri-o://4bf7bf36e3d0130cfdbc28caf002c5658e1ed940b10d8e2e726abeb5084791a2" gracePeriod=2 Dec 01 09:38:07 crc kubenswrapper[4873]: I1201 09:38:07.888325 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9l6jh" Dec 01 09:38:07 crc kubenswrapper[4873]: I1201 09:38:07.973175 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a29f28fe-2850-45a0-9b3e-eb84ef30fff3-utilities\") pod \"a29f28fe-2850-45a0-9b3e-eb84ef30fff3\" (UID: \"a29f28fe-2850-45a0-9b3e-eb84ef30fff3\") " Dec 01 09:38:07 crc kubenswrapper[4873]: I1201 09:38:07.973319 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-whhgs\" (UniqueName: \"kubernetes.io/projected/a29f28fe-2850-45a0-9b3e-eb84ef30fff3-kube-api-access-whhgs\") pod \"a29f28fe-2850-45a0-9b3e-eb84ef30fff3\" (UID: \"a29f28fe-2850-45a0-9b3e-eb84ef30fff3\") " Dec 01 09:38:07 crc kubenswrapper[4873]: I1201 09:38:07.973788 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a29f28fe-2850-45a0-9b3e-eb84ef30fff3-catalog-content\") pod \"a29f28fe-2850-45a0-9b3e-eb84ef30fff3\" (UID: \"a29f28fe-2850-45a0-9b3e-eb84ef30fff3\") " Dec 01 09:38:07 crc kubenswrapper[4873]: I1201 09:38:07.975971 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a29f28fe-2850-45a0-9b3e-eb84ef30fff3-utilities" (OuterVolumeSpecName: "utilities") pod "a29f28fe-2850-45a0-9b3e-eb84ef30fff3" (UID: "a29f28fe-2850-45a0-9b3e-eb84ef30fff3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:38:07 crc kubenswrapper[4873]: I1201 09:38:07.997343 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a29f28fe-2850-45a0-9b3e-eb84ef30fff3-kube-api-access-whhgs" (OuterVolumeSpecName: "kube-api-access-whhgs") pod "a29f28fe-2850-45a0-9b3e-eb84ef30fff3" (UID: "a29f28fe-2850-45a0-9b3e-eb84ef30fff3"). InnerVolumeSpecName "kube-api-access-whhgs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:38:08 crc kubenswrapper[4873]: I1201 09:38:08.039616 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a29f28fe-2850-45a0-9b3e-eb84ef30fff3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a29f28fe-2850-45a0-9b3e-eb84ef30fff3" (UID: "a29f28fe-2850-45a0-9b3e-eb84ef30fff3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:38:08 crc kubenswrapper[4873]: I1201 09:38:08.076456 4873 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a29f28fe-2850-45a0-9b3e-eb84ef30fff3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 09:38:08 crc kubenswrapper[4873]: I1201 09:38:08.076492 4873 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a29f28fe-2850-45a0-9b3e-eb84ef30fff3-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 09:38:08 crc kubenswrapper[4873]: I1201 09:38:08.076504 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-whhgs\" (UniqueName: \"kubernetes.io/projected/a29f28fe-2850-45a0-9b3e-eb84ef30fff3-kube-api-access-whhgs\") on node \"crc\" DevicePath \"\"" Dec 01 09:38:08 crc kubenswrapper[4873]: I1201 09:38:08.318655 4873 generic.go:334] "Generic (PLEG): container finished" podID="a29f28fe-2850-45a0-9b3e-eb84ef30fff3" containerID="4bf7bf36e3d0130cfdbc28caf002c5658e1ed940b10d8e2e726abeb5084791a2" exitCode=0 Dec 01 09:38:08 crc kubenswrapper[4873]: I1201 09:38:08.318754 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9l6jh" Dec 01 09:38:08 crc kubenswrapper[4873]: I1201 09:38:08.318785 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9l6jh" event={"ID":"a29f28fe-2850-45a0-9b3e-eb84ef30fff3","Type":"ContainerDied","Data":"4bf7bf36e3d0130cfdbc28caf002c5658e1ed940b10d8e2e726abeb5084791a2"} Dec 01 09:38:08 crc kubenswrapper[4873]: I1201 09:38:08.322609 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9l6jh" event={"ID":"a29f28fe-2850-45a0-9b3e-eb84ef30fff3","Type":"ContainerDied","Data":"b9c4478a0bc193da6c44d6d9a2830a968f72dcf0a11f2468ea6bb18563f51310"} Dec 01 09:38:08 crc kubenswrapper[4873]: I1201 09:38:08.322644 4873 scope.go:117] "RemoveContainer" containerID="4bf7bf36e3d0130cfdbc28caf002c5658e1ed940b10d8e2e726abeb5084791a2" Dec 01 09:38:08 crc kubenswrapper[4873]: I1201 09:38:08.347456 4873 scope.go:117] "RemoveContainer" containerID="8d0416829d08beb6b01c389e53dd108adb39754cc7685dfc62c9d561242d0235" Dec 01 09:38:08 crc kubenswrapper[4873]: I1201 09:38:08.369784 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9l6jh"] Dec 01 09:38:08 crc kubenswrapper[4873]: I1201 09:38:08.380441 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9l6jh"] Dec 01 09:38:08 crc kubenswrapper[4873]: I1201 09:38:08.383393 4873 scope.go:117] "RemoveContainer" containerID="eb59298f8d036d67723496770ab1fc3b339397f4240423806d6bc2122e61bcbb" Dec 01 09:38:08 crc kubenswrapper[4873]: I1201 09:38:08.435959 4873 scope.go:117] "RemoveContainer" containerID="4bf7bf36e3d0130cfdbc28caf002c5658e1ed940b10d8e2e726abeb5084791a2" Dec 01 09:38:08 crc kubenswrapper[4873]: E1201 09:38:08.436618 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4bf7bf36e3d0130cfdbc28caf002c5658e1ed940b10d8e2e726abeb5084791a2\": container with ID starting with 4bf7bf36e3d0130cfdbc28caf002c5658e1ed940b10d8e2e726abeb5084791a2 not found: ID does not exist" containerID="4bf7bf36e3d0130cfdbc28caf002c5658e1ed940b10d8e2e726abeb5084791a2" Dec 01 09:38:08 crc kubenswrapper[4873]: I1201 09:38:08.436685 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4bf7bf36e3d0130cfdbc28caf002c5658e1ed940b10d8e2e726abeb5084791a2"} err="failed to get container status \"4bf7bf36e3d0130cfdbc28caf002c5658e1ed940b10d8e2e726abeb5084791a2\": rpc error: code = NotFound desc = could not find container \"4bf7bf36e3d0130cfdbc28caf002c5658e1ed940b10d8e2e726abeb5084791a2\": container with ID starting with 4bf7bf36e3d0130cfdbc28caf002c5658e1ed940b10d8e2e726abeb5084791a2 not found: ID does not exist" Dec 01 09:38:08 crc kubenswrapper[4873]: I1201 09:38:08.436714 4873 scope.go:117] "RemoveContainer" containerID="8d0416829d08beb6b01c389e53dd108adb39754cc7685dfc62c9d561242d0235" Dec 01 09:38:08 crc kubenswrapper[4873]: E1201 09:38:08.437150 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d0416829d08beb6b01c389e53dd108adb39754cc7685dfc62c9d561242d0235\": container with ID starting with 8d0416829d08beb6b01c389e53dd108adb39754cc7685dfc62c9d561242d0235 not found: ID does not exist" containerID="8d0416829d08beb6b01c389e53dd108adb39754cc7685dfc62c9d561242d0235" Dec 01 09:38:08 crc kubenswrapper[4873]: I1201 09:38:08.437191 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d0416829d08beb6b01c389e53dd108adb39754cc7685dfc62c9d561242d0235"} err="failed to get container status \"8d0416829d08beb6b01c389e53dd108adb39754cc7685dfc62c9d561242d0235\": rpc error: code = NotFound desc = could not find container \"8d0416829d08beb6b01c389e53dd108adb39754cc7685dfc62c9d561242d0235\": container with ID starting with 8d0416829d08beb6b01c389e53dd108adb39754cc7685dfc62c9d561242d0235 not found: ID does not exist" Dec 01 09:38:08 crc kubenswrapper[4873]: I1201 09:38:08.437219 4873 scope.go:117] "RemoveContainer" containerID="eb59298f8d036d67723496770ab1fc3b339397f4240423806d6bc2122e61bcbb" Dec 01 09:38:08 crc kubenswrapper[4873]: E1201 09:38:08.437505 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb59298f8d036d67723496770ab1fc3b339397f4240423806d6bc2122e61bcbb\": container with ID starting with eb59298f8d036d67723496770ab1fc3b339397f4240423806d6bc2122e61bcbb not found: ID does not exist" containerID="eb59298f8d036d67723496770ab1fc3b339397f4240423806d6bc2122e61bcbb" Dec 01 09:38:08 crc kubenswrapper[4873]: I1201 09:38:08.437540 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb59298f8d036d67723496770ab1fc3b339397f4240423806d6bc2122e61bcbb"} err="failed to get container status \"eb59298f8d036d67723496770ab1fc3b339397f4240423806d6bc2122e61bcbb\": rpc error: code = NotFound desc = could not find container \"eb59298f8d036d67723496770ab1fc3b339397f4240423806d6bc2122e61bcbb\": container with ID starting with eb59298f8d036d67723496770ab1fc3b339397f4240423806d6bc2122e61bcbb not found: ID does not exist" Dec 01 09:38:08 crc kubenswrapper[4873]: I1201 09:38:08.442674 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a29f28fe-2850-45a0-9b3e-eb84ef30fff3" path="/var/lib/kubelet/pods/a29f28fe-2850-45a0-9b3e-eb84ef30fff3/volumes" Dec 01 09:38:10 crc kubenswrapper[4873]: I1201 09:38:10.575639 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8wrmm"] Dec 01 09:38:10 crc kubenswrapper[4873]: E1201 09:38:10.576696 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a29f28fe-2850-45a0-9b3e-eb84ef30fff3" containerName="extract-content" Dec 01 09:38:10 crc kubenswrapper[4873]: I1201 09:38:10.576713 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="a29f28fe-2850-45a0-9b3e-eb84ef30fff3" containerName="extract-content" Dec 01 09:38:10 crc kubenswrapper[4873]: E1201 09:38:10.576725 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a29f28fe-2850-45a0-9b3e-eb84ef30fff3" containerName="registry-server" Dec 01 09:38:10 crc kubenswrapper[4873]: I1201 09:38:10.576732 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="a29f28fe-2850-45a0-9b3e-eb84ef30fff3" containerName="registry-server" Dec 01 09:38:10 crc kubenswrapper[4873]: E1201 09:38:10.576770 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a29f28fe-2850-45a0-9b3e-eb84ef30fff3" containerName="extract-utilities" Dec 01 09:38:10 crc kubenswrapper[4873]: I1201 09:38:10.576777 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="a29f28fe-2850-45a0-9b3e-eb84ef30fff3" containerName="extract-utilities" Dec 01 09:38:10 crc kubenswrapper[4873]: I1201 09:38:10.576971 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="a29f28fe-2850-45a0-9b3e-eb84ef30fff3" containerName="registry-server" Dec 01 09:38:10 crc kubenswrapper[4873]: I1201 09:38:10.578499 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8wrmm" Dec 01 09:38:10 crc kubenswrapper[4873]: I1201 09:38:10.587519 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8wrmm"] Dec 01 09:38:10 crc kubenswrapper[4873]: I1201 09:38:10.642577 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzc4z\" (UniqueName: \"kubernetes.io/projected/c0ff48b1-208c-4922-9aa5-44cf60d3c8a4-kube-api-access-lzc4z\") pod \"redhat-marketplace-8wrmm\" (UID: \"c0ff48b1-208c-4922-9aa5-44cf60d3c8a4\") " pod="openshift-marketplace/redhat-marketplace-8wrmm" Dec 01 09:38:10 crc kubenswrapper[4873]: I1201 09:38:10.642867 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0ff48b1-208c-4922-9aa5-44cf60d3c8a4-catalog-content\") pod \"redhat-marketplace-8wrmm\" (UID: \"c0ff48b1-208c-4922-9aa5-44cf60d3c8a4\") " pod="openshift-marketplace/redhat-marketplace-8wrmm" Dec 01 09:38:10 crc kubenswrapper[4873]: I1201 09:38:10.643031 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0ff48b1-208c-4922-9aa5-44cf60d3c8a4-utilities\") pod \"redhat-marketplace-8wrmm\" (UID: \"c0ff48b1-208c-4922-9aa5-44cf60d3c8a4\") " pod="openshift-marketplace/redhat-marketplace-8wrmm" Dec 01 09:38:10 crc kubenswrapper[4873]: I1201 09:38:10.745771 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0ff48b1-208c-4922-9aa5-44cf60d3c8a4-catalog-content\") pod \"redhat-marketplace-8wrmm\" (UID: \"c0ff48b1-208c-4922-9aa5-44cf60d3c8a4\") " pod="openshift-marketplace/redhat-marketplace-8wrmm" Dec 01 09:38:10 crc kubenswrapper[4873]: I1201 09:38:10.745848 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0ff48b1-208c-4922-9aa5-44cf60d3c8a4-utilities\") pod \"redhat-marketplace-8wrmm\" (UID: \"c0ff48b1-208c-4922-9aa5-44cf60d3c8a4\") " pod="openshift-marketplace/redhat-marketplace-8wrmm" Dec 01 09:38:10 crc kubenswrapper[4873]: I1201 09:38:10.745979 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzc4z\" (UniqueName: \"kubernetes.io/projected/c0ff48b1-208c-4922-9aa5-44cf60d3c8a4-kube-api-access-lzc4z\") pod \"redhat-marketplace-8wrmm\" (UID: \"c0ff48b1-208c-4922-9aa5-44cf60d3c8a4\") " pod="openshift-marketplace/redhat-marketplace-8wrmm" Dec 01 09:38:10 crc kubenswrapper[4873]: I1201 09:38:10.746428 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0ff48b1-208c-4922-9aa5-44cf60d3c8a4-utilities\") pod \"redhat-marketplace-8wrmm\" (UID: \"c0ff48b1-208c-4922-9aa5-44cf60d3c8a4\") " pod="openshift-marketplace/redhat-marketplace-8wrmm" Dec 01 09:38:10 crc kubenswrapper[4873]: I1201 09:38:10.746650 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0ff48b1-208c-4922-9aa5-44cf60d3c8a4-catalog-content\") pod \"redhat-marketplace-8wrmm\" (UID: \"c0ff48b1-208c-4922-9aa5-44cf60d3c8a4\") " pod="openshift-marketplace/redhat-marketplace-8wrmm" Dec 01 09:38:10 crc kubenswrapper[4873]: I1201 09:38:10.769256 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzc4z\" (UniqueName: \"kubernetes.io/projected/c0ff48b1-208c-4922-9aa5-44cf60d3c8a4-kube-api-access-lzc4z\") pod \"redhat-marketplace-8wrmm\" (UID: \"c0ff48b1-208c-4922-9aa5-44cf60d3c8a4\") " pod="openshift-marketplace/redhat-marketplace-8wrmm" Dec 01 09:38:10 crc kubenswrapper[4873]: I1201 09:38:10.906312 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8wrmm" Dec 01 09:38:11 crc kubenswrapper[4873]: I1201 09:38:11.395648 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8wrmm"] Dec 01 09:38:12 crc kubenswrapper[4873]: I1201 09:38:12.362029 4873 generic.go:334] "Generic (PLEG): container finished" podID="c0ff48b1-208c-4922-9aa5-44cf60d3c8a4" containerID="3c125ff3b0f16130a73bb1ea0fdb4f89ace80792158963e475106fcf4a23ab5f" exitCode=0 Dec 01 09:38:12 crc kubenswrapper[4873]: I1201 09:38:12.362133 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8wrmm" event={"ID":"c0ff48b1-208c-4922-9aa5-44cf60d3c8a4","Type":"ContainerDied","Data":"3c125ff3b0f16130a73bb1ea0fdb4f89ace80792158963e475106fcf4a23ab5f"} Dec 01 09:38:12 crc kubenswrapper[4873]: I1201 09:38:12.362350 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8wrmm" event={"ID":"c0ff48b1-208c-4922-9aa5-44cf60d3c8a4","Type":"ContainerStarted","Data":"e3483cd47a74dc2d615a762affb6002c3da1ee694e96e7c30ec7ad8cfe0a35aa"} Dec 01 09:38:13 crc kubenswrapper[4873]: I1201 09:38:13.375329 4873 generic.go:334] "Generic (PLEG): container finished" podID="c0ff48b1-208c-4922-9aa5-44cf60d3c8a4" containerID="d242443e7ca93079081b871f4e8625771845345edf3966f7793465a4181c1100" exitCode=0 Dec 01 09:38:13 crc kubenswrapper[4873]: I1201 09:38:13.375714 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8wrmm" event={"ID":"c0ff48b1-208c-4922-9aa5-44cf60d3c8a4","Type":"ContainerDied","Data":"d242443e7ca93079081b871f4e8625771845345edf3966f7793465a4181c1100"} Dec 01 09:38:14 crc kubenswrapper[4873]: I1201 09:38:14.389897 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8wrmm" event={"ID":"c0ff48b1-208c-4922-9aa5-44cf60d3c8a4","Type":"ContainerStarted","Data":"d2285a67344ed4a4b465ebfa81c63f2d2d1cb06566f6875457dd1e68d956c5c7"} Dec 01 09:38:14 crc kubenswrapper[4873]: I1201 09:38:14.428380 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8wrmm" podStartSLOduration=2.95535721 podStartE2EDuration="4.428354807s" podCreationTimestamp="2025-12-01 09:38:10 +0000 UTC" firstStartedPulling="2025-12-01 09:38:12.36682507 +0000 UTC m=+3468.268933609" lastFinishedPulling="2025-12-01 09:38:13.839822667 +0000 UTC m=+3469.741931206" observedRunningTime="2025-12-01 09:38:14.420193652 +0000 UTC m=+3470.322302201" watchObservedRunningTime="2025-12-01 09:38:14.428354807 +0000 UTC m=+3470.330463356" Dec 01 09:38:20 crc kubenswrapper[4873]: I1201 09:38:20.906993 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8wrmm" Dec 01 09:38:20 crc kubenswrapper[4873]: I1201 09:38:20.908208 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8wrmm" Dec 01 09:38:20 crc kubenswrapper[4873]: I1201 09:38:20.971760 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8wrmm" Dec 01 09:38:21 crc kubenswrapper[4873]: I1201 09:38:21.621035 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8wrmm" Dec 01 09:38:21 crc kubenswrapper[4873]: I1201 09:38:21.694201 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8wrmm"] Dec 01 09:38:23 crc kubenswrapper[4873]: I1201 09:38:23.588158 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8wrmm" podUID="c0ff48b1-208c-4922-9aa5-44cf60d3c8a4" containerName="registry-server" containerID="cri-o://d2285a67344ed4a4b465ebfa81c63f2d2d1cb06566f6875457dd1e68d956c5c7" gracePeriod=2 Dec 01 09:38:23 crc kubenswrapper[4873]: E1201 09:38:23.769985 4873 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc0ff48b1_208c_4922_9aa5_44cf60d3c8a4.slice/crio-d2285a67344ed4a4b465ebfa81c63f2d2d1cb06566f6875457dd1e68d956c5c7.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc0ff48b1_208c_4922_9aa5_44cf60d3c8a4.slice/crio-conmon-d2285a67344ed4a4b465ebfa81c63f2d2d1cb06566f6875457dd1e68d956c5c7.scope\": RecentStats: unable to find data in memory cache]" Dec 01 09:38:24 crc kubenswrapper[4873]: I1201 09:38:24.148457 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8wrmm" Dec 01 09:38:24 crc kubenswrapper[4873]: I1201 09:38:24.342964 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0ff48b1-208c-4922-9aa5-44cf60d3c8a4-catalog-content\") pod \"c0ff48b1-208c-4922-9aa5-44cf60d3c8a4\" (UID: \"c0ff48b1-208c-4922-9aa5-44cf60d3c8a4\") " Dec 01 09:38:24 crc kubenswrapper[4873]: I1201 09:38:24.349172 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0ff48b1-208c-4922-9aa5-44cf60d3c8a4-utilities\") pod \"c0ff48b1-208c-4922-9aa5-44cf60d3c8a4\" (UID: \"c0ff48b1-208c-4922-9aa5-44cf60d3c8a4\") " Dec 01 09:38:24 crc kubenswrapper[4873]: I1201 09:38:24.349376 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c0ff48b1-208c-4922-9aa5-44cf60d3c8a4-utilities" (OuterVolumeSpecName: "utilities") pod "c0ff48b1-208c-4922-9aa5-44cf60d3c8a4" (UID: "c0ff48b1-208c-4922-9aa5-44cf60d3c8a4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:38:24 crc kubenswrapper[4873]: I1201 09:38:24.350336 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzc4z\" (UniqueName: \"kubernetes.io/projected/c0ff48b1-208c-4922-9aa5-44cf60d3c8a4-kube-api-access-lzc4z\") pod \"c0ff48b1-208c-4922-9aa5-44cf60d3c8a4\" (UID: \"c0ff48b1-208c-4922-9aa5-44cf60d3c8a4\") " Dec 01 09:38:24 crc kubenswrapper[4873]: I1201 09:38:24.352776 4873 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0ff48b1-208c-4922-9aa5-44cf60d3c8a4-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 09:38:24 crc kubenswrapper[4873]: I1201 09:38:24.363517 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0ff48b1-208c-4922-9aa5-44cf60d3c8a4-kube-api-access-lzc4z" (OuterVolumeSpecName: "kube-api-access-lzc4z") pod "c0ff48b1-208c-4922-9aa5-44cf60d3c8a4" (UID: "c0ff48b1-208c-4922-9aa5-44cf60d3c8a4"). InnerVolumeSpecName "kube-api-access-lzc4z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:38:24 crc kubenswrapper[4873]: I1201 09:38:24.373267 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c0ff48b1-208c-4922-9aa5-44cf60d3c8a4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c0ff48b1-208c-4922-9aa5-44cf60d3c8a4" (UID: "c0ff48b1-208c-4922-9aa5-44cf60d3c8a4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:38:24 crc kubenswrapper[4873]: I1201 09:38:24.456795 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzc4z\" (UniqueName: \"kubernetes.io/projected/c0ff48b1-208c-4922-9aa5-44cf60d3c8a4-kube-api-access-lzc4z\") on node \"crc\" DevicePath \"\"" Dec 01 09:38:24 crc kubenswrapper[4873]: I1201 09:38:24.456912 4873 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0ff48b1-208c-4922-9aa5-44cf60d3c8a4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 09:38:24 crc kubenswrapper[4873]: I1201 09:38:24.603709 4873 generic.go:334] "Generic (PLEG): container finished" podID="c0ff48b1-208c-4922-9aa5-44cf60d3c8a4" containerID="d2285a67344ed4a4b465ebfa81c63f2d2d1cb06566f6875457dd1e68d956c5c7" exitCode=0 Dec 01 09:38:24 crc kubenswrapper[4873]: I1201 09:38:24.603784 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8wrmm" event={"ID":"c0ff48b1-208c-4922-9aa5-44cf60d3c8a4","Type":"ContainerDied","Data":"d2285a67344ed4a4b465ebfa81c63f2d2d1cb06566f6875457dd1e68d956c5c7"} Dec 01 09:38:24 crc kubenswrapper[4873]: I1201 09:38:24.603834 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8wrmm" Dec 01 09:38:24 crc kubenswrapper[4873]: I1201 09:38:24.603864 4873 scope.go:117] "RemoveContainer" containerID="d2285a67344ed4a4b465ebfa81c63f2d2d1cb06566f6875457dd1e68d956c5c7" Dec 01 09:38:24 crc kubenswrapper[4873]: I1201 09:38:24.603844 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8wrmm" event={"ID":"c0ff48b1-208c-4922-9aa5-44cf60d3c8a4","Type":"ContainerDied","Data":"e3483cd47a74dc2d615a762affb6002c3da1ee694e96e7c30ec7ad8cfe0a35aa"} Dec 01 09:38:24 crc kubenswrapper[4873]: I1201 09:38:24.633909 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8wrmm"] Dec 01 09:38:24 crc kubenswrapper[4873]: I1201 09:38:24.641854 4873 scope.go:117] "RemoveContainer" containerID="d242443e7ca93079081b871f4e8625771845345edf3966f7793465a4181c1100" Dec 01 09:38:24 crc kubenswrapper[4873]: I1201 09:38:24.643762 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8wrmm"] Dec 01 09:38:24 crc kubenswrapper[4873]: I1201 09:38:24.662120 4873 scope.go:117] "RemoveContainer" containerID="3c125ff3b0f16130a73bb1ea0fdb4f89ace80792158963e475106fcf4a23ab5f" Dec 01 09:38:24 crc kubenswrapper[4873]: I1201 09:38:24.709568 4873 scope.go:117] "RemoveContainer" containerID="d2285a67344ed4a4b465ebfa81c63f2d2d1cb06566f6875457dd1e68d956c5c7" Dec 01 09:38:24 crc kubenswrapper[4873]: E1201 09:38:24.710115 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2285a67344ed4a4b465ebfa81c63f2d2d1cb06566f6875457dd1e68d956c5c7\": container with ID starting with d2285a67344ed4a4b465ebfa81c63f2d2d1cb06566f6875457dd1e68d956c5c7 not found: ID does not exist" containerID="d2285a67344ed4a4b465ebfa81c63f2d2d1cb06566f6875457dd1e68d956c5c7" Dec 01 09:38:24 crc kubenswrapper[4873]: I1201 09:38:24.710168 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2285a67344ed4a4b465ebfa81c63f2d2d1cb06566f6875457dd1e68d956c5c7"} err="failed to get container status \"d2285a67344ed4a4b465ebfa81c63f2d2d1cb06566f6875457dd1e68d956c5c7\": rpc error: code = NotFound desc = could not find container \"d2285a67344ed4a4b465ebfa81c63f2d2d1cb06566f6875457dd1e68d956c5c7\": container with ID starting with d2285a67344ed4a4b465ebfa81c63f2d2d1cb06566f6875457dd1e68d956c5c7 not found: ID does not exist" Dec 01 09:38:24 crc kubenswrapper[4873]: I1201 09:38:24.710211 4873 scope.go:117] "RemoveContainer" containerID="d242443e7ca93079081b871f4e8625771845345edf3966f7793465a4181c1100" Dec 01 09:38:24 crc kubenswrapper[4873]: E1201 09:38:24.711448 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d242443e7ca93079081b871f4e8625771845345edf3966f7793465a4181c1100\": container with ID starting with d242443e7ca93079081b871f4e8625771845345edf3966f7793465a4181c1100 not found: ID does not exist" containerID="d242443e7ca93079081b871f4e8625771845345edf3966f7793465a4181c1100" Dec 01 09:38:24 crc kubenswrapper[4873]: I1201 09:38:24.711490 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d242443e7ca93079081b871f4e8625771845345edf3966f7793465a4181c1100"} err="failed to get container status \"d242443e7ca93079081b871f4e8625771845345edf3966f7793465a4181c1100\": rpc error: code = NotFound desc = could not find container \"d242443e7ca93079081b871f4e8625771845345edf3966f7793465a4181c1100\": container with ID starting with d242443e7ca93079081b871f4e8625771845345edf3966f7793465a4181c1100 not found: ID does not exist" Dec 01 09:38:24 crc kubenswrapper[4873]: I1201 09:38:24.711512 4873 scope.go:117] "RemoveContainer" containerID="3c125ff3b0f16130a73bb1ea0fdb4f89ace80792158963e475106fcf4a23ab5f" Dec 01 09:38:24 crc kubenswrapper[4873]: E1201 09:38:24.712074 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c125ff3b0f16130a73bb1ea0fdb4f89ace80792158963e475106fcf4a23ab5f\": container with ID starting with 3c125ff3b0f16130a73bb1ea0fdb4f89ace80792158963e475106fcf4a23ab5f not found: ID does not exist" containerID="3c125ff3b0f16130a73bb1ea0fdb4f89ace80792158963e475106fcf4a23ab5f" Dec 01 09:38:24 crc kubenswrapper[4873]: I1201 09:38:24.712252 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c125ff3b0f16130a73bb1ea0fdb4f89ace80792158963e475106fcf4a23ab5f"} err="failed to get container status \"3c125ff3b0f16130a73bb1ea0fdb4f89ace80792158963e475106fcf4a23ab5f\": rpc error: code = NotFound desc = could not find container \"3c125ff3b0f16130a73bb1ea0fdb4f89ace80792158963e475106fcf4a23ab5f\": container with ID starting with 3c125ff3b0f16130a73bb1ea0fdb4f89ace80792158963e475106fcf4a23ab5f not found: ID does not exist" Dec 01 09:38:25 crc kubenswrapper[4873]: I1201 09:38:25.619120 4873 generic.go:334] "Generic (PLEG): container finished" podID="cd15a962-a485-48ad-b0af-d7a3ec0e4ab1" containerID="098e20bc0ab355e1003b658643d1f17730f91014cf149b5d9058a091734d8dd6" exitCode=0 Dec 01 09:38:25 crc kubenswrapper[4873]: I1201 09:38:25.619425 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1","Type":"ContainerDied","Data":"098e20bc0ab355e1003b658643d1f17730f91014cf149b5d9058a091734d8dd6"} Dec 01 09:38:26 crc kubenswrapper[4873]: I1201 09:38:26.443799 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0ff48b1-208c-4922-9aa5-44cf60d3c8a4" path="/var/lib/kubelet/pods/c0ff48b1-208c-4922-9aa5-44cf60d3c8a4/volumes" Dec 01 09:38:27 crc kubenswrapper[4873]: I1201 09:38:27.144537 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 01 09:38:27 crc kubenswrapper[4873]: I1201 09:38:27.333536 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-ssh-key\") pod \"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1\" (UID: \"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1\") " Dec 01 09:38:27 crc kubenswrapper[4873]: I1201 09:38:27.333644 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-openstack-config-secret\") pod \"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1\" (UID: \"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1\") " Dec 01 09:38:27 crc kubenswrapper[4873]: I1201 09:38:27.333701 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-958mm\" (UniqueName: \"kubernetes.io/projected/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-kube-api-access-958mm\") pod \"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1\" (UID: \"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1\") " Dec 01 09:38:27 crc kubenswrapper[4873]: I1201 09:38:27.333798 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-config-data\") pod \"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1\" (UID: \"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1\") " Dec 01 09:38:27 crc kubenswrapper[4873]: I1201 09:38:27.333924 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-test-operator-ephemeral-workdir\") pod \"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1\" (UID: \"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1\") " Dec 01 09:38:27 crc kubenswrapper[4873]: I1201 09:38:27.334071 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-test-operator-ephemeral-temporary\") pod \"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1\" (UID: \"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1\") " Dec 01 09:38:27 crc kubenswrapper[4873]: I1201 09:38:27.334101 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-ca-certs\") pod \"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1\" (UID: \"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1\") " Dec 01 09:38:27 crc kubenswrapper[4873]: I1201 09:38:27.334167 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1\" (UID: \"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1\") " Dec 01 09:38:27 crc kubenswrapper[4873]: I1201 09:38:27.334189 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-openstack-config\") pod \"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1\" (UID: \"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1\") " Dec 01 09:38:27 crc kubenswrapper[4873]: I1201 09:38:27.338404 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "cd15a962-a485-48ad-b0af-d7a3ec0e4ab1" (UID: "cd15a962-a485-48ad-b0af-d7a3ec0e4ab1"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:38:27 crc kubenswrapper[4873]: I1201 09:38:27.339145 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-config-data" (OuterVolumeSpecName: "config-data") pod "cd15a962-a485-48ad-b0af-d7a3ec0e4ab1" (UID: "cd15a962-a485-48ad-b0af-d7a3ec0e4ab1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:38:27 crc kubenswrapper[4873]: I1201 09:38:27.341906 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "cd15a962-a485-48ad-b0af-d7a3ec0e4ab1" (UID: "cd15a962-a485-48ad-b0af-d7a3ec0e4ab1"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:38:27 crc kubenswrapper[4873]: I1201 09:38:27.342628 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "test-operator-logs") pod "cd15a962-a485-48ad-b0af-d7a3ec0e4ab1" (UID: "cd15a962-a485-48ad-b0af-d7a3ec0e4ab1"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:38:27 crc kubenswrapper[4873]: I1201 09:38:27.344347 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-kube-api-access-958mm" (OuterVolumeSpecName: "kube-api-access-958mm") pod "cd15a962-a485-48ad-b0af-d7a3ec0e4ab1" (UID: "cd15a962-a485-48ad-b0af-d7a3ec0e4ab1"). InnerVolumeSpecName "kube-api-access-958mm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:38:27 crc kubenswrapper[4873]: I1201 09:38:27.374923 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "cd15a962-a485-48ad-b0af-d7a3ec0e4ab1" (UID: "cd15a962-a485-48ad-b0af-d7a3ec0e4ab1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:38:27 crc kubenswrapper[4873]: I1201 09:38:27.382952 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "cd15a962-a485-48ad-b0af-d7a3ec0e4ab1" (UID: "cd15a962-a485-48ad-b0af-d7a3ec0e4ab1"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:38:27 crc kubenswrapper[4873]: I1201 09:38:27.386889 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "cd15a962-a485-48ad-b0af-d7a3ec0e4ab1" (UID: "cd15a962-a485-48ad-b0af-d7a3ec0e4ab1"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:38:27 crc kubenswrapper[4873]: I1201 09:38:27.399636 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "cd15a962-a485-48ad-b0af-d7a3ec0e4ab1" (UID: "cd15a962-a485-48ad-b0af-d7a3ec0e4ab1"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:38:27 crc kubenswrapper[4873]: I1201 09:38:27.436984 4873 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:38:27 crc kubenswrapper[4873]: I1201 09:38:27.437135 4873 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 01 09:38:27 crc kubenswrapper[4873]: I1201 09:38:27.437156 4873 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 01 09:38:27 crc kubenswrapper[4873]: I1201 09:38:27.437176 4873 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 01 09:38:27 crc kubenswrapper[4873]: I1201 09:38:27.437212 4873 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Dec 01 09:38:27 crc kubenswrapper[4873]: I1201 09:38:27.437226 4873 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 01 09:38:27 crc kubenswrapper[4873]: I1201 09:38:27.437240 4873 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 09:38:27 crc kubenswrapper[4873]: I1201 09:38:27.437252 4873 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 01 09:38:27 crc kubenswrapper[4873]: I1201 09:38:27.437264 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-958mm\" (UniqueName: \"kubernetes.io/projected/cd15a962-a485-48ad-b0af-d7a3ec0e4ab1-kube-api-access-958mm\") on node \"crc\" DevicePath \"\"" Dec 01 09:38:27 crc kubenswrapper[4873]: I1201 09:38:27.469390 4873 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Dec 01 09:38:27 crc kubenswrapper[4873]: I1201 09:38:27.544592 4873 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:38:27 crc kubenswrapper[4873]: I1201 09:38:27.649469 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"cd15a962-a485-48ad-b0af-d7a3ec0e4ab1","Type":"ContainerDied","Data":"59b498de32d7198342495c719c447abce031a7c6509fe1755f6f3c7946822ed9"} Dec 01 09:38:27 crc kubenswrapper[4873]: I1201 09:38:27.649517 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59b498de32d7198342495c719c447abce031a7c6509fe1755f6f3c7946822ed9" Dec 01 09:38:27 crc kubenswrapper[4873]: I1201 09:38:27.649614 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 01 09:38:33 crc kubenswrapper[4873]: I1201 09:38:33.402138 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 01 09:38:33 crc kubenswrapper[4873]: E1201 09:38:33.403332 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0ff48b1-208c-4922-9aa5-44cf60d3c8a4" containerName="registry-server" Dec 01 09:38:33 crc kubenswrapper[4873]: I1201 09:38:33.403348 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0ff48b1-208c-4922-9aa5-44cf60d3c8a4" containerName="registry-server" Dec 01 09:38:33 crc kubenswrapper[4873]: E1201 09:38:33.403392 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd15a962-a485-48ad-b0af-d7a3ec0e4ab1" containerName="tempest-tests-tempest-tests-runner" Dec 01 09:38:33 crc kubenswrapper[4873]: I1201 09:38:33.403399 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd15a962-a485-48ad-b0af-d7a3ec0e4ab1" containerName="tempest-tests-tempest-tests-runner" Dec 01 09:38:33 crc kubenswrapper[4873]: E1201 09:38:33.403428 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0ff48b1-208c-4922-9aa5-44cf60d3c8a4" containerName="extract-content" Dec 01 09:38:33 crc kubenswrapper[4873]: I1201 09:38:33.403435 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0ff48b1-208c-4922-9aa5-44cf60d3c8a4" containerName="extract-content" Dec 01 09:38:33 crc kubenswrapper[4873]: E1201 09:38:33.403470 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0ff48b1-208c-4922-9aa5-44cf60d3c8a4" containerName="extract-utilities" Dec 01 09:38:33 crc kubenswrapper[4873]: I1201 09:38:33.403478 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0ff48b1-208c-4922-9aa5-44cf60d3c8a4" containerName="extract-utilities" Dec 01 09:38:33 crc kubenswrapper[4873]: I1201 09:38:33.403791 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd15a962-a485-48ad-b0af-d7a3ec0e4ab1" containerName="tempest-tests-tempest-tests-runner" Dec 01 09:38:33 crc kubenswrapper[4873]: I1201 09:38:33.403817 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0ff48b1-208c-4922-9aa5-44cf60d3c8a4" containerName="registry-server" Dec 01 09:38:33 crc kubenswrapper[4873]: I1201 09:38:33.405397 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 01 09:38:33 crc kubenswrapper[4873]: I1201 09:38:33.411462 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-pqxv2" Dec 01 09:38:33 crc kubenswrapper[4873]: I1201 09:38:33.412190 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 01 09:38:33 crc kubenswrapper[4873]: I1201 09:38:33.607596 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f210543d-a2a6-4746-b279-04fe37ee1b76\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 01 09:38:33 crc kubenswrapper[4873]: I1201 09:38:33.607769 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pk7r\" (UniqueName: \"kubernetes.io/projected/f210543d-a2a6-4746-b279-04fe37ee1b76-kube-api-access-6pk7r\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f210543d-a2a6-4746-b279-04fe37ee1b76\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 01 09:38:33 crc kubenswrapper[4873]: I1201 09:38:33.710489 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f210543d-a2a6-4746-b279-04fe37ee1b76\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 01 09:38:33 crc kubenswrapper[4873]: I1201 09:38:33.710600 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pk7r\" (UniqueName: \"kubernetes.io/projected/f210543d-a2a6-4746-b279-04fe37ee1b76-kube-api-access-6pk7r\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f210543d-a2a6-4746-b279-04fe37ee1b76\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 01 09:38:33 crc kubenswrapper[4873]: I1201 09:38:33.711287 4873 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f210543d-a2a6-4746-b279-04fe37ee1b76\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 01 09:38:33 crc kubenswrapper[4873]: I1201 09:38:33.743122 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pk7r\" (UniqueName: \"kubernetes.io/projected/f210543d-a2a6-4746-b279-04fe37ee1b76-kube-api-access-6pk7r\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f210543d-a2a6-4746-b279-04fe37ee1b76\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 01 09:38:33 crc kubenswrapper[4873]: I1201 09:38:33.745610 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f210543d-a2a6-4746-b279-04fe37ee1b76\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 01 09:38:34 crc kubenswrapper[4873]: I1201 09:38:34.043155 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 01 09:38:34 crc kubenswrapper[4873]: W1201 09:38:34.539458 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf210543d_a2a6_4746_b279_04fe37ee1b76.slice/crio-b33cf646702c6e7e70c56dc945c5fc0c16c3f3dea8c0c644e3aa412769b1522f WatchSource:0}: Error finding container b33cf646702c6e7e70c56dc945c5fc0c16c3f3dea8c0c644e3aa412769b1522f: Status 404 returned error can't find the container with id b33cf646702c6e7e70c56dc945c5fc0c16c3f3dea8c0c644e3aa412769b1522f Dec 01 09:38:34 crc kubenswrapper[4873]: I1201 09:38:34.540112 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 01 09:38:34 crc kubenswrapper[4873]: I1201 09:38:34.725873 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"f210543d-a2a6-4746-b279-04fe37ee1b76","Type":"ContainerStarted","Data":"b33cf646702c6e7e70c56dc945c5fc0c16c3f3dea8c0c644e3aa412769b1522f"} Dec 01 09:38:36 crc kubenswrapper[4873]: I1201 09:38:36.750607 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"f210543d-a2a6-4746-b279-04fe37ee1b76","Type":"ContainerStarted","Data":"8d1a0c9fd16f728f3b27235b9496d50171e70940043a62e6327fa8dc596bc31f"} Dec 01 09:38:36 crc kubenswrapper[4873]: I1201 09:38:36.772804 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.572583169 podStartE2EDuration="3.772775819s" podCreationTimestamp="2025-12-01 09:38:33 +0000 UTC" firstStartedPulling="2025-12-01 09:38:34.541906167 +0000 UTC m=+3490.444014716" lastFinishedPulling="2025-12-01 09:38:35.742098827 +0000 UTC m=+3491.644207366" observedRunningTime="2025-12-01 09:38:36.764348937 +0000 UTC m=+3492.666457476" watchObservedRunningTime="2025-12-01 09:38:36.772775819 +0000 UTC m=+3492.674884358" Dec 01 09:39:00 crc kubenswrapper[4873]: I1201 09:39:00.698368 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-2kbfg/must-gather-x85kc"] Dec 01 09:39:00 crc kubenswrapper[4873]: I1201 09:39:00.700838 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2kbfg/must-gather-x85kc" Dec 01 09:39:00 crc kubenswrapper[4873]: I1201 09:39:00.705594 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-2kbfg"/"kube-root-ca.crt" Dec 01 09:39:00 crc kubenswrapper[4873]: I1201 09:39:00.709136 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-2kbfg"/"openshift-service-ca.crt" Dec 01 09:39:00 crc kubenswrapper[4873]: I1201 09:39:00.709446 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-2kbfg"/"default-dockercfg-hpzsc" Dec 01 09:39:00 crc kubenswrapper[4873]: I1201 09:39:00.732550 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-2kbfg/must-gather-x85kc"] Dec 01 09:39:00 crc kubenswrapper[4873]: I1201 09:39:00.745003 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mttpd\" (UniqueName: \"kubernetes.io/projected/87fffa2f-8ebb-413f-a56d-cfe1d67eea39-kube-api-access-mttpd\") pod \"must-gather-x85kc\" (UID: \"87fffa2f-8ebb-413f-a56d-cfe1d67eea39\") " pod="openshift-must-gather-2kbfg/must-gather-x85kc" Dec 01 09:39:00 crc kubenswrapper[4873]: I1201 09:39:00.745612 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/87fffa2f-8ebb-413f-a56d-cfe1d67eea39-must-gather-output\") pod \"must-gather-x85kc\" (UID: \"87fffa2f-8ebb-413f-a56d-cfe1d67eea39\") " pod="openshift-must-gather-2kbfg/must-gather-x85kc" Dec 01 09:39:00 crc kubenswrapper[4873]: I1201 09:39:00.849813 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/87fffa2f-8ebb-413f-a56d-cfe1d67eea39-must-gather-output\") pod \"must-gather-x85kc\" (UID: \"87fffa2f-8ebb-413f-a56d-cfe1d67eea39\") " pod="openshift-must-gather-2kbfg/must-gather-x85kc" Dec 01 09:39:00 crc kubenswrapper[4873]: I1201 09:39:00.849965 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mttpd\" (UniqueName: \"kubernetes.io/projected/87fffa2f-8ebb-413f-a56d-cfe1d67eea39-kube-api-access-mttpd\") pod \"must-gather-x85kc\" (UID: \"87fffa2f-8ebb-413f-a56d-cfe1d67eea39\") " pod="openshift-must-gather-2kbfg/must-gather-x85kc" Dec 01 09:39:00 crc kubenswrapper[4873]: I1201 09:39:00.851725 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/87fffa2f-8ebb-413f-a56d-cfe1d67eea39-must-gather-output\") pod \"must-gather-x85kc\" (UID: \"87fffa2f-8ebb-413f-a56d-cfe1d67eea39\") " pod="openshift-must-gather-2kbfg/must-gather-x85kc" Dec 01 09:39:00 crc kubenswrapper[4873]: I1201 09:39:00.892636 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mttpd\" (UniqueName: \"kubernetes.io/projected/87fffa2f-8ebb-413f-a56d-cfe1d67eea39-kube-api-access-mttpd\") pod \"must-gather-x85kc\" (UID: \"87fffa2f-8ebb-413f-a56d-cfe1d67eea39\") " pod="openshift-must-gather-2kbfg/must-gather-x85kc" Dec 01 09:39:01 crc kubenswrapper[4873]: I1201 09:39:01.054063 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2kbfg/must-gather-x85kc" Dec 01 09:39:01 crc kubenswrapper[4873]: I1201 09:39:01.059704 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:39:01 crc kubenswrapper[4873]: I1201 09:39:01.059783 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:39:01 crc kubenswrapper[4873]: I1201 09:39:01.602741 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-2kbfg/must-gather-x85kc"] Dec 01 09:39:01 crc kubenswrapper[4873]: W1201 09:39:01.611198 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod87fffa2f_8ebb_413f_a56d_cfe1d67eea39.slice/crio-7f6c443d95ea21668e740e38b6224ecf436514316cf136d4d7b1ab576ca12d58 WatchSource:0}: Error finding container 7f6c443d95ea21668e740e38b6224ecf436514316cf136d4d7b1ab576ca12d58: Status 404 returned error can't find the container with id 7f6c443d95ea21668e740e38b6224ecf436514316cf136d4d7b1ab576ca12d58 Dec 01 09:39:02 crc kubenswrapper[4873]: I1201 09:39:02.060081 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2kbfg/must-gather-x85kc" event={"ID":"87fffa2f-8ebb-413f-a56d-cfe1d67eea39","Type":"ContainerStarted","Data":"7f6c443d95ea21668e740e38b6224ecf436514316cf136d4d7b1ab576ca12d58"} Dec 01 09:39:06 crc kubenswrapper[4873]: I1201 09:39:06.102439 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2kbfg/must-gather-x85kc" event={"ID":"87fffa2f-8ebb-413f-a56d-cfe1d67eea39","Type":"ContainerStarted","Data":"3b94b41bfb72d83a88c0bcb88544f1492a872fadc16ba4d4c3e873ec15d64d3a"} Dec 01 09:39:07 crc kubenswrapper[4873]: I1201 09:39:07.127967 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2kbfg/must-gather-x85kc" event={"ID":"87fffa2f-8ebb-413f-a56d-cfe1d67eea39","Type":"ContainerStarted","Data":"a6cf1033c5faf41459c9638ee3912415e6056d18d22703fb9f627730fbfc3793"} Dec 01 09:39:07 crc kubenswrapper[4873]: I1201 09:39:07.152919 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-2kbfg/must-gather-x85kc" podStartSLOduration=3.066200352 podStartE2EDuration="7.152898515s" podCreationTimestamp="2025-12-01 09:39:00 +0000 UTC" firstStartedPulling="2025-12-01 09:39:01.613826431 +0000 UTC m=+3517.515934970" lastFinishedPulling="2025-12-01 09:39:05.700524594 +0000 UTC m=+3521.602633133" observedRunningTime="2025-12-01 09:39:07.147582482 +0000 UTC m=+3523.049691021" watchObservedRunningTime="2025-12-01 09:39:07.152898515 +0000 UTC m=+3523.055007054" Dec 01 09:39:10 crc kubenswrapper[4873]: I1201 09:39:10.361516 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-2kbfg/crc-debug-7s8pr"] Dec 01 09:39:10 crc kubenswrapper[4873]: I1201 09:39:10.364802 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2kbfg/crc-debug-7s8pr" Dec 01 09:39:10 crc kubenswrapper[4873]: I1201 09:39:10.391878 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmnbj\" (UniqueName: \"kubernetes.io/projected/250aed85-9559-4d19-beda-85868b1ed372-kube-api-access-hmnbj\") pod \"crc-debug-7s8pr\" (UID: \"250aed85-9559-4d19-beda-85868b1ed372\") " pod="openshift-must-gather-2kbfg/crc-debug-7s8pr" Dec 01 09:39:10 crc kubenswrapper[4873]: I1201 09:39:10.392160 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/250aed85-9559-4d19-beda-85868b1ed372-host\") pod \"crc-debug-7s8pr\" (UID: \"250aed85-9559-4d19-beda-85868b1ed372\") " pod="openshift-must-gather-2kbfg/crc-debug-7s8pr" Dec 01 09:39:10 crc kubenswrapper[4873]: I1201 09:39:10.494714 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmnbj\" (UniqueName: \"kubernetes.io/projected/250aed85-9559-4d19-beda-85868b1ed372-kube-api-access-hmnbj\") pod \"crc-debug-7s8pr\" (UID: \"250aed85-9559-4d19-beda-85868b1ed372\") " pod="openshift-must-gather-2kbfg/crc-debug-7s8pr" Dec 01 09:39:10 crc kubenswrapper[4873]: I1201 09:39:10.494895 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/250aed85-9559-4d19-beda-85868b1ed372-host\") pod \"crc-debug-7s8pr\" (UID: \"250aed85-9559-4d19-beda-85868b1ed372\") " pod="openshift-must-gather-2kbfg/crc-debug-7s8pr" Dec 01 09:39:10 crc kubenswrapper[4873]: I1201 09:39:10.495124 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/250aed85-9559-4d19-beda-85868b1ed372-host\") pod \"crc-debug-7s8pr\" (UID: \"250aed85-9559-4d19-beda-85868b1ed372\") " pod="openshift-must-gather-2kbfg/crc-debug-7s8pr" Dec 01 09:39:10 crc kubenswrapper[4873]: I1201 09:39:10.546150 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmnbj\" (UniqueName: \"kubernetes.io/projected/250aed85-9559-4d19-beda-85868b1ed372-kube-api-access-hmnbj\") pod \"crc-debug-7s8pr\" (UID: \"250aed85-9559-4d19-beda-85868b1ed372\") " pod="openshift-must-gather-2kbfg/crc-debug-7s8pr" Dec 01 09:39:10 crc kubenswrapper[4873]: I1201 09:39:10.685382 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2kbfg/crc-debug-7s8pr" Dec 01 09:39:10 crc kubenswrapper[4873]: W1201 09:39:10.721478 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod250aed85_9559_4d19_beda_85868b1ed372.slice/crio-626b688a2d52b831be26e210f71ac85156bf38b65bd3b2b2a46adc3e72f65719 WatchSource:0}: Error finding container 626b688a2d52b831be26e210f71ac85156bf38b65bd3b2b2a46adc3e72f65719: Status 404 returned error can't find the container with id 626b688a2d52b831be26e210f71ac85156bf38b65bd3b2b2a46adc3e72f65719 Dec 01 09:39:11 crc kubenswrapper[4873]: I1201 09:39:11.168397 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2kbfg/crc-debug-7s8pr" event={"ID":"250aed85-9559-4d19-beda-85868b1ed372","Type":"ContainerStarted","Data":"626b688a2d52b831be26e210f71ac85156bf38b65bd3b2b2a46adc3e72f65719"} Dec 01 09:39:24 crc kubenswrapper[4873]: I1201 09:39:24.325869 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2kbfg/crc-debug-7s8pr" event={"ID":"250aed85-9559-4d19-beda-85868b1ed372","Type":"ContainerStarted","Data":"1d366b6cea0d6f8316e95277ce12c45b342842e883466bf11357e84802083e3f"} Dec 01 09:39:31 crc kubenswrapper[4873]: I1201 09:39:31.058961 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:39:31 crc kubenswrapper[4873]: I1201 09:39:31.060487 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:39:41 crc kubenswrapper[4873]: I1201 09:39:41.490343 4873 generic.go:334] "Generic (PLEG): container finished" podID="250aed85-9559-4d19-beda-85868b1ed372" containerID="1d366b6cea0d6f8316e95277ce12c45b342842e883466bf11357e84802083e3f" exitCode=0 Dec 01 09:39:41 crc kubenswrapper[4873]: I1201 09:39:41.490424 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2kbfg/crc-debug-7s8pr" event={"ID":"250aed85-9559-4d19-beda-85868b1ed372","Type":"ContainerDied","Data":"1d366b6cea0d6f8316e95277ce12c45b342842e883466bf11357e84802083e3f"} Dec 01 09:39:42 crc kubenswrapper[4873]: I1201 09:39:42.643267 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2kbfg/crc-debug-7s8pr" Dec 01 09:39:42 crc kubenswrapper[4873]: I1201 09:39:42.670313 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/250aed85-9559-4d19-beda-85868b1ed372-host\") pod \"250aed85-9559-4d19-beda-85868b1ed372\" (UID: \"250aed85-9559-4d19-beda-85868b1ed372\") " Dec 01 09:39:42 crc kubenswrapper[4873]: I1201 09:39:42.670723 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hmnbj\" (UniqueName: \"kubernetes.io/projected/250aed85-9559-4d19-beda-85868b1ed372-kube-api-access-hmnbj\") pod \"250aed85-9559-4d19-beda-85868b1ed372\" (UID: \"250aed85-9559-4d19-beda-85868b1ed372\") " Dec 01 09:39:42 crc kubenswrapper[4873]: I1201 09:39:42.672554 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/250aed85-9559-4d19-beda-85868b1ed372-host" (OuterVolumeSpecName: "host") pod "250aed85-9559-4d19-beda-85868b1ed372" (UID: "250aed85-9559-4d19-beda-85868b1ed372"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:39:42 crc kubenswrapper[4873]: I1201 09:39:42.678988 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/250aed85-9559-4d19-beda-85868b1ed372-kube-api-access-hmnbj" (OuterVolumeSpecName: "kube-api-access-hmnbj") pod "250aed85-9559-4d19-beda-85868b1ed372" (UID: "250aed85-9559-4d19-beda-85868b1ed372"). InnerVolumeSpecName "kube-api-access-hmnbj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:39:42 crc kubenswrapper[4873]: I1201 09:39:42.685168 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-2kbfg/crc-debug-7s8pr"] Dec 01 09:39:42 crc kubenswrapper[4873]: I1201 09:39:42.697144 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-2kbfg/crc-debug-7s8pr"] Dec 01 09:39:42 crc kubenswrapper[4873]: I1201 09:39:42.773417 4873 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/250aed85-9559-4d19-beda-85868b1ed372-host\") on node \"crc\" DevicePath \"\"" Dec 01 09:39:42 crc kubenswrapper[4873]: I1201 09:39:42.773457 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hmnbj\" (UniqueName: \"kubernetes.io/projected/250aed85-9559-4d19-beda-85868b1ed372-kube-api-access-hmnbj\") on node \"crc\" DevicePath \"\"" Dec 01 09:39:43 crc kubenswrapper[4873]: I1201 09:39:43.518756 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="626b688a2d52b831be26e210f71ac85156bf38b65bd3b2b2a46adc3e72f65719" Dec 01 09:39:43 crc kubenswrapper[4873]: I1201 09:39:43.518871 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2kbfg/crc-debug-7s8pr" Dec 01 09:39:43 crc kubenswrapper[4873]: I1201 09:39:43.893861 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-2kbfg/crc-debug-xg76w"] Dec 01 09:39:43 crc kubenswrapper[4873]: E1201 09:39:43.894404 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="250aed85-9559-4d19-beda-85868b1ed372" containerName="container-00" Dec 01 09:39:43 crc kubenswrapper[4873]: I1201 09:39:43.894421 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="250aed85-9559-4d19-beda-85868b1ed372" containerName="container-00" Dec 01 09:39:43 crc kubenswrapper[4873]: I1201 09:39:43.894676 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="250aed85-9559-4d19-beda-85868b1ed372" containerName="container-00" Dec 01 09:39:43 crc kubenswrapper[4873]: I1201 09:39:43.895543 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2kbfg/crc-debug-xg76w" Dec 01 09:39:43 crc kubenswrapper[4873]: I1201 09:39:43.999331 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cdmt\" (UniqueName: \"kubernetes.io/projected/cc2fa4b7-9fee-4899-81ce-d32fb755d317-kube-api-access-4cdmt\") pod \"crc-debug-xg76w\" (UID: \"cc2fa4b7-9fee-4899-81ce-d32fb755d317\") " pod="openshift-must-gather-2kbfg/crc-debug-xg76w" Dec 01 09:39:43 crc kubenswrapper[4873]: I1201 09:39:43.999540 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cc2fa4b7-9fee-4899-81ce-d32fb755d317-host\") pod \"crc-debug-xg76w\" (UID: \"cc2fa4b7-9fee-4899-81ce-d32fb755d317\") " pod="openshift-must-gather-2kbfg/crc-debug-xg76w" Dec 01 09:39:44 crc kubenswrapper[4873]: I1201 09:39:44.101606 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cc2fa4b7-9fee-4899-81ce-d32fb755d317-host\") pod \"crc-debug-xg76w\" (UID: \"cc2fa4b7-9fee-4899-81ce-d32fb755d317\") " pod="openshift-must-gather-2kbfg/crc-debug-xg76w" Dec 01 09:39:44 crc kubenswrapper[4873]: I1201 09:39:44.101701 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cdmt\" (UniqueName: \"kubernetes.io/projected/cc2fa4b7-9fee-4899-81ce-d32fb755d317-kube-api-access-4cdmt\") pod \"crc-debug-xg76w\" (UID: \"cc2fa4b7-9fee-4899-81ce-d32fb755d317\") " pod="openshift-must-gather-2kbfg/crc-debug-xg76w" Dec 01 09:39:44 crc kubenswrapper[4873]: I1201 09:39:44.102190 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cc2fa4b7-9fee-4899-81ce-d32fb755d317-host\") pod \"crc-debug-xg76w\" (UID: \"cc2fa4b7-9fee-4899-81ce-d32fb755d317\") " pod="openshift-must-gather-2kbfg/crc-debug-xg76w" Dec 01 09:39:44 crc kubenswrapper[4873]: I1201 09:39:44.126422 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cdmt\" (UniqueName: \"kubernetes.io/projected/cc2fa4b7-9fee-4899-81ce-d32fb755d317-kube-api-access-4cdmt\") pod \"crc-debug-xg76w\" (UID: \"cc2fa4b7-9fee-4899-81ce-d32fb755d317\") " pod="openshift-must-gather-2kbfg/crc-debug-xg76w" Dec 01 09:39:44 crc kubenswrapper[4873]: I1201 09:39:44.213155 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2kbfg/crc-debug-xg76w" Dec 01 09:39:44 crc kubenswrapper[4873]: I1201 09:39:44.468100 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="250aed85-9559-4d19-beda-85868b1ed372" path="/var/lib/kubelet/pods/250aed85-9559-4d19-beda-85868b1ed372/volumes" Dec 01 09:39:44 crc kubenswrapper[4873]: I1201 09:39:44.529715 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2kbfg/crc-debug-xg76w" event={"ID":"cc2fa4b7-9fee-4899-81ce-d32fb755d317","Type":"ContainerStarted","Data":"5f32451391a904d861ed3c2f3e00b0187fe674153b567027ae6c5f37551956f1"} Dec 01 09:39:45 crc kubenswrapper[4873]: I1201 09:39:45.541741 4873 generic.go:334] "Generic (PLEG): container finished" podID="cc2fa4b7-9fee-4899-81ce-d32fb755d317" containerID="f5e5825d7c2d3938a584f70ea3f41dbcea5fa91e7581901ec1c03cf589c405ed" exitCode=1 Dec 01 09:39:45 crc kubenswrapper[4873]: I1201 09:39:45.541852 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2kbfg/crc-debug-xg76w" event={"ID":"cc2fa4b7-9fee-4899-81ce-d32fb755d317","Type":"ContainerDied","Data":"f5e5825d7c2d3938a584f70ea3f41dbcea5fa91e7581901ec1c03cf589c405ed"} Dec 01 09:39:45 crc kubenswrapper[4873]: I1201 09:39:45.591447 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-2kbfg/crc-debug-xg76w"] Dec 01 09:39:45 crc kubenswrapper[4873]: I1201 09:39:45.606398 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-2kbfg/crc-debug-xg76w"] Dec 01 09:39:46 crc kubenswrapper[4873]: I1201 09:39:46.679134 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2kbfg/crc-debug-xg76w" Dec 01 09:39:46 crc kubenswrapper[4873]: I1201 09:39:46.764843 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cc2fa4b7-9fee-4899-81ce-d32fb755d317-host\") pod \"cc2fa4b7-9fee-4899-81ce-d32fb755d317\" (UID: \"cc2fa4b7-9fee-4899-81ce-d32fb755d317\") " Dec 01 09:39:46 crc kubenswrapper[4873]: I1201 09:39:46.764981 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc2fa4b7-9fee-4899-81ce-d32fb755d317-host" (OuterVolumeSpecName: "host") pod "cc2fa4b7-9fee-4899-81ce-d32fb755d317" (UID: "cc2fa4b7-9fee-4899-81ce-d32fb755d317"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:39:46 crc kubenswrapper[4873]: I1201 09:39:46.765008 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4cdmt\" (UniqueName: \"kubernetes.io/projected/cc2fa4b7-9fee-4899-81ce-d32fb755d317-kube-api-access-4cdmt\") pod \"cc2fa4b7-9fee-4899-81ce-d32fb755d317\" (UID: \"cc2fa4b7-9fee-4899-81ce-d32fb755d317\") " Dec 01 09:39:46 crc kubenswrapper[4873]: I1201 09:39:46.765760 4873 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cc2fa4b7-9fee-4899-81ce-d32fb755d317-host\") on node \"crc\" DevicePath \"\"" Dec 01 09:39:46 crc kubenswrapper[4873]: I1201 09:39:46.771152 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc2fa4b7-9fee-4899-81ce-d32fb755d317-kube-api-access-4cdmt" (OuterVolumeSpecName: "kube-api-access-4cdmt") pod "cc2fa4b7-9fee-4899-81ce-d32fb755d317" (UID: "cc2fa4b7-9fee-4899-81ce-d32fb755d317"). InnerVolumeSpecName "kube-api-access-4cdmt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:39:46 crc kubenswrapper[4873]: I1201 09:39:46.868106 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4cdmt\" (UniqueName: \"kubernetes.io/projected/cc2fa4b7-9fee-4899-81ce-d32fb755d317-kube-api-access-4cdmt\") on node \"crc\" DevicePath \"\"" Dec 01 09:39:47 crc kubenswrapper[4873]: I1201 09:39:47.565340 4873 scope.go:117] "RemoveContainer" containerID="f5e5825d7c2d3938a584f70ea3f41dbcea5fa91e7581901ec1c03cf589c405ed" Dec 01 09:39:47 crc kubenswrapper[4873]: I1201 09:39:47.565379 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2kbfg/crc-debug-xg76w" Dec 01 09:39:48 crc kubenswrapper[4873]: I1201 09:39:48.441887 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc2fa4b7-9fee-4899-81ce-d32fb755d317" path="/var/lib/kubelet/pods/cc2fa4b7-9fee-4899-81ce-d32fb755d317/volumes" Dec 01 09:40:01 crc kubenswrapper[4873]: I1201 09:40:01.058759 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:40:01 crc kubenswrapper[4873]: I1201 09:40:01.059430 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:40:01 crc kubenswrapper[4873]: I1201 09:40:01.059494 4873 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" Dec 01 09:40:01 crc kubenswrapper[4873]: I1201 09:40:01.060177 4873 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c101914cc0d7462afe57653cc6d66f5335bbd707b5b5acec3231122718624d93"} pod="openshift-machine-config-operator/machine-config-daemon-scwpp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 09:40:01 crc kubenswrapper[4873]: I1201 09:40:01.060239 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" containerID="cri-o://c101914cc0d7462afe57653cc6d66f5335bbd707b5b5acec3231122718624d93" gracePeriod=600 Dec 01 09:40:01 crc kubenswrapper[4873]: E1201 09:40:01.186913 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:40:01 crc kubenswrapper[4873]: I1201 09:40:01.711437 4873 generic.go:334] "Generic (PLEG): container finished" podID="fef7b114-0e07-402d-a37b-315c36011f4b" containerID="c101914cc0d7462afe57653cc6d66f5335bbd707b5b5acec3231122718624d93" exitCode=0 Dec 01 09:40:01 crc kubenswrapper[4873]: I1201 09:40:01.711497 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" event={"ID":"fef7b114-0e07-402d-a37b-315c36011f4b","Type":"ContainerDied","Data":"c101914cc0d7462afe57653cc6d66f5335bbd707b5b5acec3231122718624d93"} Dec 01 09:40:01 crc kubenswrapper[4873]: I1201 09:40:01.711549 4873 scope.go:117] "RemoveContainer" containerID="0b72d3c91d2960798896e9520de55a4bb91b7db511fab82dd6f7e66083a742f2" Dec 01 09:40:01 crc kubenswrapper[4873]: I1201 09:40:01.712612 4873 scope.go:117] "RemoveContainer" containerID="c101914cc0d7462afe57653cc6d66f5335bbd707b5b5acec3231122718624d93" Dec 01 09:40:01 crc kubenswrapper[4873]: E1201 09:40:01.712959 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:40:13 crc kubenswrapper[4873]: I1201 09:40:13.430771 4873 scope.go:117] "RemoveContainer" containerID="c101914cc0d7462afe57653cc6d66f5335bbd707b5b5acec3231122718624d93" Dec 01 09:40:13 crc kubenswrapper[4873]: E1201 09:40:13.431690 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:40:26 crc kubenswrapper[4873]: I1201 09:40:26.990645 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-d4t9d"] Dec 01 09:40:26 crc kubenswrapper[4873]: E1201 09:40:26.991833 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc2fa4b7-9fee-4899-81ce-d32fb755d317" containerName="container-00" Dec 01 09:40:26 crc kubenswrapper[4873]: I1201 09:40:26.991855 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc2fa4b7-9fee-4899-81ce-d32fb755d317" containerName="container-00" Dec 01 09:40:26 crc kubenswrapper[4873]: I1201 09:40:26.992130 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc2fa4b7-9fee-4899-81ce-d32fb755d317" containerName="container-00" Dec 01 09:40:26 crc kubenswrapper[4873]: I1201 09:40:26.994248 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d4t9d" Dec 01 09:40:27 crc kubenswrapper[4873]: I1201 09:40:27.003763 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d4t9d"] Dec 01 09:40:27 crc kubenswrapper[4873]: I1201 09:40:27.046462 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18da147b-c2cb-4217-887f-d22da4e3e38c-utilities\") pod \"redhat-operators-d4t9d\" (UID: \"18da147b-c2cb-4217-887f-d22da4e3e38c\") " pod="openshift-marketplace/redhat-operators-d4t9d" Dec 01 09:40:27 crc kubenswrapper[4873]: I1201 09:40:27.046565 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nd2s9\" (UniqueName: \"kubernetes.io/projected/18da147b-c2cb-4217-887f-d22da4e3e38c-kube-api-access-nd2s9\") pod \"redhat-operators-d4t9d\" (UID: \"18da147b-c2cb-4217-887f-d22da4e3e38c\") " pod="openshift-marketplace/redhat-operators-d4t9d" Dec 01 09:40:27 crc kubenswrapper[4873]: I1201 09:40:27.046595 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18da147b-c2cb-4217-887f-d22da4e3e38c-catalog-content\") pod \"redhat-operators-d4t9d\" (UID: \"18da147b-c2cb-4217-887f-d22da4e3e38c\") " pod="openshift-marketplace/redhat-operators-d4t9d" Dec 01 09:40:27 crc kubenswrapper[4873]: I1201 09:40:27.148825 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18da147b-c2cb-4217-887f-d22da4e3e38c-utilities\") pod \"redhat-operators-d4t9d\" (UID: \"18da147b-c2cb-4217-887f-d22da4e3e38c\") " pod="openshift-marketplace/redhat-operators-d4t9d" Dec 01 09:40:27 crc kubenswrapper[4873]: I1201 09:40:27.149252 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nd2s9\" (UniqueName: \"kubernetes.io/projected/18da147b-c2cb-4217-887f-d22da4e3e38c-kube-api-access-nd2s9\") pod \"redhat-operators-d4t9d\" (UID: \"18da147b-c2cb-4217-887f-d22da4e3e38c\") " pod="openshift-marketplace/redhat-operators-d4t9d" Dec 01 09:40:27 crc kubenswrapper[4873]: I1201 09:40:27.149284 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18da147b-c2cb-4217-887f-d22da4e3e38c-catalog-content\") pod \"redhat-operators-d4t9d\" (UID: \"18da147b-c2cb-4217-887f-d22da4e3e38c\") " pod="openshift-marketplace/redhat-operators-d4t9d" Dec 01 09:40:27 crc kubenswrapper[4873]: I1201 09:40:27.149642 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18da147b-c2cb-4217-887f-d22da4e3e38c-catalog-content\") pod \"redhat-operators-d4t9d\" (UID: \"18da147b-c2cb-4217-887f-d22da4e3e38c\") " pod="openshift-marketplace/redhat-operators-d4t9d" Dec 01 09:40:27 crc kubenswrapper[4873]: I1201 09:40:27.149728 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18da147b-c2cb-4217-887f-d22da4e3e38c-utilities\") pod \"redhat-operators-d4t9d\" (UID: \"18da147b-c2cb-4217-887f-d22da4e3e38c\") " pod="openshift-marketplace/redhat-operators-d4t9d" Dec 01 09:40:27 crc kubenswrapper[4873]: I1201 09:40:27.178772 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nd2s9\" (UniqueName: \"kubernetes.io/projected/18da147b-c2cb-4217-887f-d22da4e3e38c-kube-api-access-nd2s9\") pod \"redhat-operators-d4t9d\" (UID: \"18da147b-c2cb-4217-887f-d22da4e3e38c\") " pod="openshift-marketplace/redhat-operators-d4t9d" Dec 01 09:40:27 crc kubenswrapper[4873]: I1201 09:40:27.348147 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d4t9d" Dec 01 09:40:27 crc kubenswrapper[4873]: I1201 09:40:27.429832 4873 scope.go:117] "RemoveContainer" containerID="c101914cc0d7462afe57653cc6d66f5335bbd707b5b5acec3231122718624d93" Dec 01 09:40:27 crc kubenswrapper[4873]: E1201 09:40:27.430534 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:40:27 crc kubenswrapper[4873]: I1201 09:40:27.906770 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d4t9d"] Dec 01 09:40:28 crc kubenswrapper[4873]: I1201 09:40:28.042233 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d4t9d" event={"ID":"18da147b-c2cb-4217-887f-d22da4e3e38c","Type":"ContainerStarted","Data":"6e260d2c83d7f64ce197217cbd622f6741c350e6be4e4b5eb9aad39c62136ed8"} Dec 01 09:40:29 crc kubenswrapper[4873]: I1201 09:40:29.055549 4873 generic.go:334] "Generic (PLEG): container finished" podID="18da147b-c2cb-4217-887f-d22da4e3e38c" containerID="97b743a4a46ed1ca1322aacd4e67a403f465ed732c27ebfc5d5b160d90c893bd" exitCode=0 Dec 01 09:40:29 crc kubenswrapper[4873]: I1201 09:40:29.055710 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d4t9d" event={"ID":"18da147b-c2cb-4217-887f-d22da4e3e38c","Type":"ContainerDied","Data":"97b743a4a46ed1ca1322aacd4e67a403f465ed732c27ebfc5d5b160d90c893bd"} Dec 01 09:40:30 crc kubenswrapper[4873]: I1201 09:40:30.182246 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qkqh2"] Dec 01 09:40:30 crc kubenswrapper[4873]: I1201 09:40:30.184848 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qkqh2" Dec 01 09:40:30 crc kubenswrapper[4873]: I1201 09:40:30.197321 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qkqh2"] Dec 01 09:40:30 crc kubenswrapper[4873]: I1201 09:40:30.220407 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjmss\" (UniqueName: \"kubernetes.io/projected/2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1-kube-api-access-jjmss\") pod \"certified-operators-qkqh2\" (UID: \"2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1\") " pod="openshift-marketplace/certified-operators-qkqh2" Dec 01 09:40:30 crc kubenswrapper[4873]: I1201 09:40:30.220702 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1-utilities\") pod \"certified-operators-qkqh2\" (UID: \"2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1\") " pod="openshift-marketplace/certified-operators-qkqh2" Dec 01 09:40:30 crc kubenswrapper[4873]: I1201 09:40:30.220977 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1-catalog-content\") pod \"certified-operators-qkqh2\" (UID: \"2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1\") " pod="openshift-marketplace/certified-operators-qkqh2" Dec 01 09:40:30 crc kubenswrapper[4873]: I1201 09:40:30.324195 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjmss\" (UniqueName: \"kubernetes.io/projected/2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1-kube-api-access-jjmss\") pod \"certified-operators-qkqh2\" (UID: \"2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1\") " pod="openshift-marketplace/certified-operators-qkqh2" Dec 01 09:40:30 crc kubenswrapper[4873]: I1201 09:40:30.324313 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1-utilities\") pod \"certified-operators-qkqh2\" (UID: \"2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1\") " pod="openshift-marketplace/certified-operators-qkqh2" Dec 01 09:40:30 crc kubenswrapper[4873]: I1201 09:40:30.324380 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1-catalog-content\") pod \"certified-operators-qkqh2\" (UID: \"2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1\") " pod="openshift-marketplace/certified-operators-qkqh2" Dec 01 09:40:30 crc kubenswrapper[4873]: I1201 09:40:30.325139 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1-catalog-content\") pod \"certified-operators-qkqh2\" (UID: \"2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1\") " pod="openshift-marketplace/certified-operators-qkqh2" Dec 01 09:40:30 crc kubenswrapper[4873]: I1201 09:40:30.325601 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1-utilities\") pod \"certified-operators-qkqh2\" (UID: \"2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1\") " pod="openshift-marketplace/certified-operators-qkqh2" Dec 01 09:40:30 crc kubenswrapper[4873]: I1201 09:40:30.357721 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjmss\" (UniqueName: \"kubernetes.io/projected/2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1-kube-api-access-jjmss\") pod \"certified-operators-qkqh2\" (UID: \"2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1\") " pod="openshift-marketplace/certified-operators-qkqh2" Dec 01 09:40:30 crc kubenswrapper[4873]: I1201 09:40:30.507071 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qkqh2" Dec 01 09:40:31 crc kubenswrapper[4873]: I1201 09:40:31.080744 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d4t9d" event={"ID":"18da147b-c2cb-4217-887f-d22da4e3e38c","Type":"ContainerStarted","Data":"bf4f3b5ef0f530150093f471395e5c5eb6229762f8fbcfb72a7a6f498704e482"} Dec 01 09:40:31 crc kubenswrapper[4873]: I1201 09:40:31.241242 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qkqh2"] Dec 01 09:40:32 crc kubenswrapper[4873]: I1201 09:40:32.095216 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qkqh2" event={"ID":"2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1","Type":"ContainerStarted","Data":"0a018ff61b8604185775effc16fe4775e7dfcb947bb3a867618bad9c04056a0a"} Dec 01 09:40:32 crc kubenswrapper[4873]: I1201 09:40:32.095671 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qkqh2" event={"ID":"2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1","Type":"ContainerStarted","Data":"f44cd4fcedbe5747af3213af8090a928443a7268cc99eec8e34a7e1cc05ab2b1"} Dec 01 09:40:33 crc kubenswrapper[4873]: I1201 09:40:33.108479 4873 generic.go:334] "Generic (PLEG): container finished" podID="18da147b-c2cb-4217-887f-d22da4e3e38c" containerID="bf4f3b5ef0f530150093f471395e5c5eb6229762f8fbcfb72a7a6f498704e482" exitCode=0 Dec 01 09:40:33 crc kubenswrapper[4873]: I1201 09:40:33.108557 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d4t9d" event={"ID":"18da147b-c2cb-4217-887f-d22da4e3e38c","Type":"ContainerDied","Data":"bf4f3b5ef0f530150093f471395e5c5eb6229762f8fbcfb72a7a6f498704e482"} Dec 01 09:40:34 crc kubenswrapper[4873]: I1201 09:40:34.121169 4873 generic.go:334] "Generic (PLEG): container finished" podID="2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1" containerID="0a018ff61b8604185775effc16fe4775e7dfcb947bb3a867618bad9c04056a0a" exitCode=0 Dec 01 09:40:34 crc kubenswrapper[4873]: I1201 09:40:34.121252 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qkqh2" event={"ID":"2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1","Type":"ContainerDied","Data":"0a018ff61b8604185775effc16fe4775e7dfcb947bb3a867618bad9c04056a0a"} Dec 01 09:40:35 crc kubenswrapper[4873]: I1201 09:40:35.134998 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d4t9d" event={"ID":"18da147b-c2cb-4217-887f-d22da4e3e38c","Type":"ContainerStarted","Data":"d20539d5bec37c17589b2ae5430c6a8767b60ba135c4471731e2e27ae12868a3"} Dec 01 09:40:35 crc kubenswrapper[4873]: I1201 09:40:35.167242 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-d4t9d" podStartSLOduration=3.856249929 podStartE2EDuration="9.167220197s" podCreationTimestamp="2025-12-01 09:40:26 +0000 UTC" firstStartedPulling="2025-12-01 09:40:29.0582608 +0000 UTC m=+3604.960369349" lastFinishedPulling="2025-12-01 09:40:34.369231078 +0000 UTC m=+3610.271339617" observedRunningTime="2025-12-01 09:40:35.161213047 +0000 UTC m=+3611.063321586" watchObservedRunningTime="2025-12-01 09:40:35.167220197 +0000 UTC m=+3611.069328736" Dec 01 09:40:36 crc kubenswrapper[4873]: I1201 09:40:36.148458 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qkqh2" event={"ID":"2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1","Type":"ContainerStarted","Data":"4754f6dbade8a2f059f7306533f6a37613eab84a5c0308a397b0726bfd811b67"} Dec 01 09:40:37 crc kubenswrapper[4873]: I1201 09:40:37.168780 4873 generic.go:334] "Generic (PLEG): container finished" podID="2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1" containerID="4754f6dbade8a2f059f7306533f6a37613eab84a5c0308a397b0726bfd811b67" exitCode=0 Dec 01 09:40:37 crc kubenswrapper[4873]: I1201 09:40:37.169141 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qkqh2" event={"ID":"2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1","Type":"ContainerDied","Data":"4754f6dbade8a2f059f7306533f6a37613eab84a5c0308a397b0726bfd811b67"} Dec 01 09:40:37 crc kubenswrapper[4873]: I1201 09:40:37.349225 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-d4t9d" Dec 01 09:40:37 crc kubenswrapper[4873]: I1201 09:40:37.349286 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-d4t9d" Dec 01 09:40:38 crc kubenswrapper[4873]: I1201 09:40:38.408403 4873 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-d4t9d" podUID="18da147b-c2cb-4217-887f-d22da4e3e38c" containerName="registry-server" probeResult="failure" output=< Dec 01 09:40:38 crc kubenswrapper[4873]: timeout: failed to connect service ":50051" within 1s Dec 01 09:40:38 crc kubenswrapper[4873]: > Dec 01 09:40:39 crc kubenswrapper[4873]: I1201 09:40:39.194739 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qkqh2" event={"ID":"2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1","Type":"ContainerStarted","Data":"d9cf361e0021f16d018de134a3b2f836dbfd7fc427e3684f89d89d167508c309"} Dec 01 09:40:39 crc kubenswrapper[4873]: I1201 09:40:39.218249 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qkqh2" podStartSLOduration=5.316126599 podStartE2EDuration="9.218221646s" podCreationTimestamp="2025-12-01 09:40:30 +0000 UTC" firstStartedPulling="2025-12-01 09:40:34.125176711 +0000 UTC m=+3610.027285250" lastFinishedPulling="2025-12-01 09:40:38.027271758 +0000 UTC m=+3613.929380297" observedRunningTime="2025-12-01 09:40:39.217237702 +0000 UTC m=+3615.119346241" watchObservedRunningTime="2025-12-01 09:40:39.218221646 +0000 UTC m=+3615.120330185" Dec 01 09:40:40 crc kubenswrapper[4873]: I1201 09:40:40.435622 4873 scope.go:117] "RemoveContainer" containerID="c101914cc0d7462afe57653cc6d66f5335bbd707b5b5acec3231122718624d93" Dec 01 09:40:40 crc kubenswrapper[4873]: E1201 09:40:40.437830 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:40:40 crc kubenswrapper[4873]: I1201 09:40:40.508069 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qkqh2" Dec 01 09:40:40 crc kubenswrapper[4873]: I1201 09:40:40.508628 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qkqh2" Dec 01 09:40:40 crc kubenswrapper[4873]: I1201 09:40:40.593368 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qkqh2" Dec 01 09:40:45 crc kubenswrapper[4873]: I1201 09:40:45.764884 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7fc6f67df6-72xpt_145a11d3-f1db-45aa-a3fe-0d4709905406/barbican-api/0.log" Dec 01 09:40:45 crc kubenswrapper[4873]: I1201 09:40:45.881224 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7fc6f67df6-72xpt_145a11d3-f1db-45aa-a3fe-0d4709905406/barbican-api-log/0.log" Dec 01 09:40:46 crc kubenswrapper[4873]: I1201 09:40:46.093794 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6674fb675b-58jdf_99f7daba-8483-4494-b783-9628f828ea49/barbican-keystone-listener/0.log" Dec 01 09:40:46 crc kubenswrapper[4873]: I1201 09:40:46.352636 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6674fb675b-58jdf_99f7daba-8483-4494-b783-9628f828ea49/barbican-keystone-listener-log/0.log" Dec 01 09:40:46 crc kubenswrapper[4873]: I1201 09:40:46.387996 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-9b48ddc7c-fn75z_4d0787b5-0da8-493b-b55d-b6e36a759f7f/barbican-worker/0.log" Dec 01 09:40:46 crc kubenswrapper[4873]: I1201 09:40:46.465082 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-9b48ddc7c-fn75z_4d0787b5-0da8-493b-b55d-b6e36a759f7f/barbican-worker-log/0.log" Dec 01 09:40:46 crc kubenswrapper[4873]: I1201 09:40:46.632362 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-wfccb_cd2a1185-cebc-4f89-88c2-63cfebe97759/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:40:47 crc kubenswrapper[4873]: I1201 09:40:47.021653 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_6ccdf302-89b3-4dc2-94c4-45ea2c2ea241/ceilometer-central-agent/0.log" Dec 01 09:40:47 crc kubenswrapper[4873]: I1201 09:40:47.088731 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_6ccdf302-89b3-4dc2-94c4-45ea2c2ea241/ceilometer-notification-agent/0.log" Dec 01 09:40:47 crc kubenswrapper[4873]: I1201 09:40:47.206167 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_6ccdf302-89b3-4dc2-94c4-45ea2c2ea241/proxy-httpd/0.log" Dec 01 09:40:47 crc kubenswrapper[4873]: I1201 09:40:47.281102 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_6ccdf302-89b3-4dc2-94c4-45ea2c2ea241/sg-core/0.log" Dec 01 09:40:47 crc kubenswrapper[4873]: I1201 09:40:47.361574 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-edpm-deployment-openstack-edpm-ipam-cz28q_d186292e-15d4-4944-b6ac-7db1fb7b85f5/ceph-client-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:40:47 crc kubenswrapper[4873]: I1201 09:40:47.414660 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-d4t9d" Dec 01 09:40:47 crc kubenswrapper[4873]: I1201 09:40:47.483001 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-d4t9d" Dec 01 09:40:47 crc kubenswrapper[4873]: I1201 09:40:47.609430 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-75xtr_b355b9ce-a737-407d-8c9e-5d9db024bb10/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:40:47 crc kubenswrapper[4873]: I1201 09:40:47.657751 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-d4t9d"] Dec 01 09:40:47 crc kubenswrapper[4873]: I1201 09:40:47.699870 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_50bc343c-5aca-46d3-a9de-31546ac7c45f/cinder-api/0.log" Dec 01 09:40:47 crc kubenswrapper[4873]: I1201 09:40:47.844353 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_50bc343c-5aca-46d3-a9de-31546ac7c45f/cinder-api-log/0.log" Dec 01 09:40:48 crc kubenswrapper[4873]: I1201 09:40:48.115145 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_3ba752bd-4b19-4ac4-9798-60de8b5a6f3e/probe/0.log" Dec 01 09:40:48 crc kubenswrapper[4873]: I1201 09:40:48.300960 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_3ba752bd-4b19-4ac4-9798-60de8b5a6f3e/cinder-backup/0.log" Dec 01 09:40:48 crc kubenswrapper[4873]: I1201 09:40:48.318584 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_4c035924-7ef3-4fc1-8652-78317f6b5c70/cinder-scheduler/0.log" Dec 01 09:40:48 crc kubenswrapper[4873]: I1201 09:40:48.441580 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_4c035924-7ef3-4fc1-8652-78317f6b5c70/probe/0.log" Dec 01 09:40:48 crc kubenswrapper[4873]: I1201 09:40:48.651374 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_96e55f05-de52-4914-9fc8-7634bff464a7/cinder-volume/0.log" Dec 01 09:40:48 crc kubenswrapper[4873]: I1201 09:40:48.680979 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_96e55f05-de52-4914-9fc8-7634bff464a7/probe/0.log" Dec 01 09:40:48 crc kubenswrapper[4873]: I1201 09:40:48.970320 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-kntht_cf256dda-e402-4a9f-bff1-fe2990f7ce72/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:40:49 crc kubenswrapper[4873]: I1201 09:40:49.027593 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-lcck8_99a70cc5-d0e6-4066-9fdb-7524bf04b7ab/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:40:49 crc kubenswrapper[4873]: I1201 09:40:49.233174 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-76b5fdb995-hdbhs_90639568-2248-4249-be34-140ed23a5d1d/init/0.log" Dec 01 09:40:49 crc kubenswrapper[4873]: I1201 09:40:49.304635 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-d4t9d" podUID="18da147b-c2cb-4217-887f-d22da4e3e38c" containerName="registry-server" containerID="cri-o://d20539d5bec37c17589b2ae5430c6a8767b60ba135c4471731e2e27ae12868a3" gracePeriod=2 Dec 01 09:40:49 crc kubenswrapper[4873]: I1201 09:40:49.598562 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-76b5fdb995-hdbhs_90639568-2248-4249-be34-140ed23a5d1d/init/0.log" Dec 01 09:40:49 crc kubenswrapper[4873]: I1201 09:40:49.728643 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_380e16b7-2a33-4591-bb99-9382512be92c/glance-httpd/0.log" Dec 01 09:40:49 crc kubenswrapper[4873]: I1201 09:40:49.736806 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-76b5fdb995-hdbhs_90639568-2248-4249-be34-140ed23a5d1d/dnsmasq-dns/0.log" Dec 01 09:40:49 crc kubenswrapper[4873]: I1201 09:40:49.855436 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_380e16b7-2a33-4591-bb99-9382512be92c/glance-log/0.log" Dec 01 09:40:50 crc kubenswrapper[4873]: I1201 09:40:50.085284 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_90a86592-ced2-4308-91fc-e9993667556a/glance-httpd/0.log" Dec 01 09:40:50 crc kubenswrapper[4873]: I1201 09:40:50.193179 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_90a86592-ced2-4308-91fc-e9993667556a/glance-log/0.log" Dec 01 09:40:50 crc kubenswrapper[4873]: I1201 09:40:50.351915 4873 generic.go:334] "Generic (PLEG): container finished" podID="18da147b-c2cb-4217-887f-d22da4e3e38c" containerID="d20539d5bec37c17589b2ae5430c6a8767b60ba135c4471731e2e27ae12868a3" exitCode=0 Dec 01 09:40:50 crc kubenswrapper[4873]: I1201 09:40:50.352005 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d4t9d" event={"ID":"18da147b-c2cb-4217-887f-d22da4e3e38c","Type":"ContainerDied","Data":"d20539d5bec37c17589b2ae5430c6a8767b60ba135c4471731e2e27ae12868a3"} Dec 01 09:40:50 crc kubenswrapper[4873]: I1201 09:40:50.495950 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d4t9d" Dec 01 09:40:50 crc kubenswrapper[4873]: I1201 09:40:50.576531 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6968fbf848-w68ch_7f8ecf06-f050-454b-8d36-e91b49847601/horizon-log/0.log" Dec 01 09:40:50 crc kubenswrapper[4873]: I1201 09:40:50.583305 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qkqh2" Dec 01 09:40:50 crc kubenswrapper[4873]: I1201 09:40:50.598261 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6968fbf848-w68ch_7f8ecf06-f050-454b-8d36-e91b49847601/horizon/0.log" Dec 01 09:40:50 crc kubenswrapper[4873]: I1201 09:40:50.659517 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nd2s9\" (UniqueName: \"kubernetes.io/projected/18da147b-c2cb-4217-887f-d22da4e3e38c-kube-api-access-nd2s9\") pod \"18da147b-c2cb-4217-887f-d22da4e3e38c\" (UID: \"18da147b-c2cb-4217-887f-d22da4e3e38c\") " Dec 01 09:40:50 crc kubenswrapper[4873]: I1201 09:40:50.659623 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18da147b-c2cb-4217-887f-d22da4e3e38c-utilities\") pod \"18da147b-c2cb-4217-887f-d22da4e3e38c\" (UID: \"18da147b-c2cb-4217-887f-d22da4e3e38c\") " Dec 01 09:40:50 crc kubenswrapper[4873]: I1201 09:40:50.659840 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18da147b-c2cb-4217-887f-d22da4e3e38c-catalog-content\") pod \"18da147b-c2cb-4217-887f-d22da4e3e38c\" (UID: \"18da147b-c2cb-4217-887f-d22da4e3e38c\") " Dec 01 09:40:50 crc kubenswrapper[4873]: I1201 09:40:50.661708 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18da147b-c2cb-4217-887f-d22da4e3e38c-utilities" (OuterVolumeSpecName: "utilities") pod "18da147b-c2cb-4217-887f-d22da4e3e38c" (UID: "18da147b-c2cb-4217-887f-d22da4e3e38c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:40:50 crc kubenswrapper[4873]: I1201 09:40:50.673687 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18da147b-c2cb-4217-887f-d22da4e3e38c-kube-api-access-nd2s9" (OuterVolumeSpecName: "kube-api-access-nd2s9") pod "18da147b-c2cb-4217-887f-d22da4e3e38c" (UID: "18da147b-c2cb-4217-887f-d22da4e3e38c"). InnerVolumeSpecName "kube-api-access-nd2s9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:40:50 crc kubenswrapper[4873]: I1201 09:40:50.719862 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj_36c9f7db-7138-48d1-9e4d-f04be409c123/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:40:50 crc kubenswrapper[4873]: I1201 09:40:50.762889 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nd2s9\" (UniqueName: \"kubernetes.io/projected/18da147b-c2cb-4217-887f-d22da4e3e38c-kube-api-access-nd2s9\") on node \"crc\" DevicePath \"\"" Dec 01 09:40:50 crc kubenswrapper[4873]: I1201 09:40:50.762936 4873 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18da147b-c2cb-4217-887f-d22da4e3e38c-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 09:40:50 crc kubenswrapper[4873]: I1201 09:40:50.771666 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18da147b-c2cb-4217-887f-d22da4e3e38c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "18da147b-c2cb-4217-887f-d22da4e3e38c" (UID: "18da147b-c2cb-4217-887f-d22da4e3e38c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:40:50 crc kubenswrapper[4873]: I1201 09:40:50.865093 4873 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18da147b-c2cb-4217-887f-d22da4e3e38c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 09:40:50 crc kubenswrapper[4873]: I1201 09:40:50.934307 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-99878_83908788-7c1f-410a-be4d-79510d6703b5/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:40:51 crc kubenswrapper[4873]: I1201 09:40:51.100146 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qkqh2"] Dec 01 09:40:51 crc kubenswrapper[4873]: I1201 09:40:51.102042 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-58576c9895-sk9tb_0857cf38-58dd-416b-92ca-6b98800ff512/keystone-api/0.log" Dec 01 09:40:51 crc kubenswrapper[4873]: I1201 09:40:51.139844 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29409661-gls8t_cf6891b3-2f06-4218-b321-5d7fac6edb7c/keystone-cron/0.log" Dec 01 09:40:51 crc kubenswrapper[4873]: I1201 09:40:51.367831 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qkqh2" podUID="2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1" containerName="registry-server" containerID="cri-o://d9cf361e0021f16d018de134a3b2f836dbfd7fc427e3684f89d89d167508c309" gracePeriod=2 Dec 01 09:40:51 crc kubenswrapper[4873]: I1201 09:40:51.367984 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d4t9d" Dec 01 09:40:51 crc kubenswrapper[4873]: I1201 09:40:51.371224 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d4t9d" event={"ID":"18da147b-c2cb-4217-887f-d22da4e3e38c","Type":"ContainerDied","Data":"6e260d2c83d7f64ce197217cbd622f6741c350e6be4e4b5eb9aad39c62136ed8"} Dec 01 09:40:51 crc kubenswrapper[4873]: I1201 09:40:51.371319 4873 scope.go:117] "RemoveContainer" containerID="d20539d5bec37c17589b2ae5430c6a8767b60ba135c4471731e2e27ae12868a3" Dec 01 09:40:51 crc kubenswrapper[4873]: I1201 09:40:51.449073 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-d4t9d"] Dec 01 09:40:51 crc kubenswrapper[4873]: I1201 09:40:51.463547 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-d4t9d"] Dec 01 09:40:51 crc kubenswrapper[4873]: I1201 09:40:51.467327 4873 scope.go:117] "RemoveContainer" containerID="bf4f3b5ef0f530150093f471395e5c5eb6229762f8fbcfb72a7a6f498704e482" Dec 01 09:40:51 crc kubenswrapper[4873]: I1201 09:40:51.591860 4873 scope.go:117] "RemoveContainer" containerID="97b743a4a46ed1ca1322aacd4e67a403f465ed732c27ebfc5d5b160d90c893bd" Dec 01 09:40:51 crc kubenswrapper[4873]: I1201 09:40:51.622098 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_84cdb02b-04a2-4456-bd2e-48d9e9a896dd/kube-state-metrics/0.log" Dec 01 09:40:51 crc kubenswrapper[4873]: I1201 09:40:51.781433 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp_328559dd-d6ca-400a-9bb1-955781e2e1ea/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:40:51 crc kubenswrapper[4873]: I1201 09:40:51.924798 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-9a52-account-create-update-vtg5d_483831f0-8b26-4845-b893-50bb7fb6ddb4/mariadb-account-create-update/0.log" Dec 01 09:40:52 crc kubenswrapper[4873]: I1201 09:40:52.085443 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qkqh2" Dec 01 09:40:52 crc kubenswrapper[4873]: I1201 09:40:52.118217 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_7744a110-0501-4606-9c09-a6442c0fefb0/manila-api/0.log" Dec 01 09:40:52 crc kubenswrapper[4873]: I1201 09:40:52.185681 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_7744a110-0501-4606-9c09-a6442c0fefb0/manila-api-log/0.log" Dec 01 09:40:52 crc kubenswrapper[4873]: I1201 09:40:52.196199 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1-catalog-content\") pod \"2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1\" (UID: \"2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1\") " Dec 01 09:40:52 crc kubenswrapper[4873]: I1201 09:40:52.196353 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1-utilities\") pod \"2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1\" (UID: \"2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1\") " Dec 01 09:40:52 crc kubenswrapper[4873]: I1201 09:40:52.196409 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jjmss\" (UniqueName: \"kubernetes.io/projected/2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1-kube-api-access-jjmss\") pod \"2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1\" (UID: \"2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1\") " Dec 01 09:40:52 crc kubenswrapper[4873]: I1201 09:40:52.197497 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1-utilities" (OuterVolumeSpecName: "utilities") pod "2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1" (UID: "2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:40:52 crc kubenswrapper[4873]: I1201 09:40:52.208194 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1-kube-api-access-jjmss" (OuterVolumeSpecName: "kube-api-access-jjmss") pod "2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1" (UID: "2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1"). InnerVolumeSpecName "kube-api-access-jjmss". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:40:52 crc kubenswrapper[4873]: I1201 09:40:52.247215 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-db-create-v7b4d_a8a98544-560c-487a-9a71-59b53a882c9b/mariadb-database-create/0.log" Dec 01 09:40:52 crc kubenswrapper[4873]: I1201 09:40:52.254903 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1" (UID: "2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:40:52 crc kubenswrapper[4873]: I1201 09:40:52.299732 4873 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 09:40:52 crc kubenswrapper[4873]: I1201 09:40:52.299824 4873 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 09:40:52 crc kubenswrapper[4873]: I1201 09:40:52.299839 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jjmss\" (UniqueName: \"kubernetes.io/projected/2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1-kube-api-access-jjmss\") on node \"crc\" DevicePath \"\"" Dec 01 09:40:52 crc kubenswrapper[4873]: I1201 09:40:52.381276 4873 generic.go:334] "Generic (PLEG): container finished" podID="2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1" containerID="d9cf361e0021f16d018de134a3b2f836dbfd7fc427e3684f89d89d167508c309" exitCode=0 Dec 01 09:40:52 crc kubenswrapper[4873]: I1201 09:40:52.381367 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qkqh2" event={"ID":"2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1","Type":"ContainerDied","Data":"d9cf361e0021f16d018de134a3b2f836dbfd7fc427e3684f89d89d167508c309"} Dec 01 09:40:52 crc kubenswrapper[4873]: I1201 09:40:52.381396 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qkqh2" Dec 01 09:40:52 crc kubenswrapper[4873]: I1201 09:40:52.381766 4873 scope.go:117] "RemoveContainer" containerID="d9cf361e0021f16d018de134a3b2f836dbfd7fc427e3684f89d89d167508c309" Dec 01 09:40:52 crc kubenswrapper[4873]: I1201 09:40:52.381740 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qkqh2" event={"ID":"2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1","Type":"ContainerDied","Data":"f44cd4fcedbe5747af3213af8090a928443a7268cc99eec8e34a7e1cc05ab2b1"} Dec 01 09:40:52 crc kubenswrapper[4873]: I1201 09:40:52.404789 4873 scope.go:117] "RemoveContainer" containerID="4754f6dbade8a2f059f7306533f6a37613eab84a5c0308a397b0726bfd811b67" Dec 01 09:40:52 crc kubenswrapper[4873]: I1201 09:40:52.429347 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qkqh2"] Dec 01 09:40:52 crc kubenswrapper[4873]: I1201 09:40:52.442992 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-db-sync-4fq4c_f07d0052-de5b-4bc4-9b7b-d65cf86edd06/manila-db-sync/0.log" Dec 01 09:40:52 crc kubenswrapper[4873]: I1201 09:40:52.444318 4873 scope.go:117] "RemoveContainer" containerID="0a018ff61b8604185775effc16fe4775e7dfcb947bb3a867618bad9c04056a0a" Dec 01 09:40:52 crc kubenswrapper[4873]: I1201 09:40:52.452168 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18da147b-c2cb-4217-887f-d22da4e3e38c" path="/var/lib/kubelet/pods/18da147b-c2cb-4217-887f-d22da4e3e38c/volumes" Dec 01 09:40:52 crc kubenswrapper[4873]: I1201 09:40:52.456747 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qkqh2"] Dec 01 09:40:52 crc kubenswrapper[4873]: I1201 09:40:52.465254 4873 scope.go:117] "RemoveContainer" containerID="d9cf361e0021f16d018de134a3b2f836dbfd7fc427e3684f89d89d167508c309" Dec 01 09:40:52 crc kubenswrapper[4873]: E1201 09:40:52.465859 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9cf361e0021f16d018de134a3b2f836dbfd7fc427e3684f89d89d167508c309\": container with ID starting with d9cf361e0021f16d018de134a3b2f836dbfd7fc427e3684f89d89d167508c309 not found: ID does not exist" containerID="d9cf361e0021f16d018de134a3b2f836dbfd7fc427e3684f89d89d167508c309" Dec 01 09:40:52 crc kubenswrapper[4873]: I1201 09:40:52.465891 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9cf361e0021f16d018de134a3b2f836dbfd7fc427e3684f89d89d167508c309"} err="failed to get container status \"d9cf361e0021f16d018de134a3b2f836dbfd7fc427e3684f89d89d167508c309\": rpc error: code = NotFound desc = could not find container \"d9cf361e0021f16d018de134a3b2f836dbfd7fc427e3684f89d89d167508c309\": container with ID starting with d9cf361e0021f16d018de134a3b2f836dbfd7fc427e3684f89d89d167508c309 not found: ID does not exist" Dec 01 09:40:52 crc kubenswrapper[4873]: I1201 09:40:52.465916 4873 scope.go:117] "RemoveContainer" containerID="4754f6dbade8a2f059f7306533f6a37613eab84a5c0308a397b0726bfd811b67" Dec 01 09:40:52 crc kubenswrapper[4873]: E1201 09:40:52.466284 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4754f6dbade8a2f059f7306533f6a37613eab84a5c0308a397b0726bfd811b67\": container with ID starting with 4754f6dbade8a2f059f7306533f6a37613eab84a5c0308a397b0726bfd811b67 not found: ID does not exist" containerID="4754f6dbade8a2f059f7306533f6a37613eab84a5c0308a397b0726bfd811b67" Dec 01 09:40:52 crc kubenswrapper[4873]: I1201 09:40:52.466304 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4754f6dbade8a2f059f7306533f6a37613eab84a5c0308a397b0726bfd811b67"} err="failed to get container status \"4754f6dbade8a2f059f7306533f6a37613eab84a5c0308a397b0726bfd811b67\": rpc error: code = NotFound desc = could not find container \"4754f6dbade8a2f059f7306533f6a37613eab84a5c0308a397b0726bfd811b67\": container with ID starting with 4754f6dbade8a2f059f7306533f6a37613eab84a5c0308a397b0726bfd811b67 not found: ID does not exist" Dec 01 09:40:52 crc kubenswrapper[4873]: I1201 09:40:52.466330 4873 scope.go:117] "RemoveContainer" containerID="0a018ff61b8604185775effc16fe4775e7dfcb947bb3a867618bad9c04056a0a" Dec 01 09:40:52 crc kubenswrapper[4873]: E1201 09:40:52.466771 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a018ff61b8604185775effc16fe4775e7dfcb947bb3a867618bad9c04056a0a\": container with ID starting with 0a018ff61b8604185775effc16fe4775e7dfcb947bb3a867618bad9c04056a0a not found: ID does not exist" containerID="0a018ff61b8604185775effc16fe4775e7dfcb947bb3a867618bad9c04056a0a" Dec 01 09:40:52 crc kubenswrapper[4873]: I1201 09:40:52.466807 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a018ff61b8604185775effc16fe4775e7dfcb947bb3a867618bad9c04056a0a"} err="failed to get container status \"0a018ff61b8604185775effc16fe4775e7dfcb947bb3a867618bad9c04056a0a\": rpc error: code = NotFound desc = could not find container \"0a018ff61b8604185775effc16fe4775e7dfcb947bb3a867618bad9c04056a0a\": container with ID starting with 0a018ff61b8604185775effc16fe4775e7dfcb947bb3a867618bad9c04056a0a not found: ID does not exist" Dec 01 09:40:52 crc kubenswrapper[4873]: I1201 09:40:52.520702 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_3f333cac-7ae8-4563-a71e-3439f5b7717d/probe/0.log" Dec 01 09:40:52 crc kubenswrapper[4873]: I1201 09:40:52.565096 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_3f333cac-7ae8-4563-a71e-3439f5b7717d/manila-scheduler/0.log" Dec 01 09:40:52 crc kubenswrapper[4873]: I1201 09:40:52.762750 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_7f96788d-8f75-490b-99b7-08096baaab34/manila-share/0.log" Dec 01 09:40:52 crc kubenswrapper[4873]: I1201 09:40:52.845216 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_7f96788d-8f75-490b-99b7-08096baaab34/probe/0.log" Dec 01 09:40:53 crc kubenswrapper[4873]: I1201 09:40:53.221841 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-f99dddf57-sr27n_7b48bbdb-ef25-4a25-a79f-d01eca3d63df/neutron-api/0.log" Dec 01 09:40:53 crc kubenswrapper[4873]: I1201 09:40:53.251779 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-f99dddf57-sr27n_7b48bbdb-ef25-4a25-a79f-d01eca3d63df/neutron-httpd/0.log" Dec 01 09:40:53 crc kubenswrapper[4873]: I1201 09:40:53.425710 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b_ed4f574a-846e-41be-88ce-56d89a392a87/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:40:53 crc kubenswrapper[4873]: I1201 09:40:53.907203 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_11dc97a8-cb16-4992-abc5-4fe5de83bac0/nova-api-log/0.log" Dec 01 09:40:53 crc kubenswrapper[4873]: I1201 09:40:53.942382 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_e18f53d4-43ea-46f1-ae6f-5eeee3ac5b5d/nova-cell0-conductor-conductor/0.log" Dec 01 09:40:54 crc kubenswrapper[4873]: I1201 09:40:54.037328 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_11dc97a8-cb16-4992-abc5-4fe5de83bac0/nova-api-api/0.log" Dec 01 09:40:54 crc kubenswrapper[4873]: I1201 09:40:54.361702 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_425ae46c-8af5-46d8-bc7c-e7056aebdd34/nova-cell1-conductor-conductor/0.log" Dec 01 09:40:54 crc kubenswrapper[4873]: I1201 09:40:54.424170 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_ae20a830-d3d8-42fc-9fff-1aabc6ac8139/nova-cell1-novncproxy-novncproxy/0.log" Dec 01 09:40:54 crc kubenswrapper[4873]: I1201 09:40:54.452152 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1" path="/var/lib/kubelet/pods/2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1/volumes" Dec 01 09:40:54 crc kubenswrapper[4873]: I1201 09:40:54.599130 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp_38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:40:54 crc kubenswrapper[4873]: I1201 09:40:54.772099 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_e54ecbbf-ac5c-4924-a396-549da1855e65/nova-metadata-log/0.log" Dec 01 09:40:55 crc kubenswrapper[4873]: I1201 09:40:55.089527 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_634610e2-dd0b-46f5-a934-e7f44249c731/nova-scheduler-scheduler/0.log" Dec 01 09:40:55 crc kubenswrapper[4873]: I1201 09:40:55.139616 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_714ffca1-ee33-4f92-a55e-29e2edf84eff/mysql-bootstrap/0.log" Dec 01 09:40:55 crc kubenswrapper[4873]: I1201 09:40:55.438265 4873 scope.go:117] "RemoveContainer" containerID="c101914cc0d7462afe57653cc6d66f5335bbd707b5b5acec3231122718624d93" Dec 01 09:40:55 crc kubenswrapper[4873]: E1201 09:40:55.438537 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:40:55 crc kubenswrapper[4873]: I1201 09:40:55.658542 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_714ffca1-ee33-4f92-a55e-29e2edf84eff/mysql-bootstrap/0.log" Dec 01 09:40:55 crc kubenswrapper[4873]: I1201 09:40:55.705741 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_714ffca1-ee33-4f92-a55e-29e2edf84eff/galera/0.log" Dec 01 09:40:55 crc kubenswrapper[4873]: I1201 09:40:55.962766 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_250cbd69-c69a-475c-8973-bc98caf4b264/mysql-bootstrap/0.log" Dec 01 09:40:56 crc kubenswrapper[4873]: I1201 09:40:56.141345 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_250cbd69-c69a-475c-8973-bc98caf4b264/mysql-bootstrap/0.log" Dec 01 09:40:56 crc kubenswrapper[4873]: I1201 09:40:56.250747 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_250cbd69-c69a-475c-8973-bc98caf4b264/galera/0.log" Dec 01 09:40:56 crc kubenswrapper[4873]: I1201 09:40:56.451668 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_e54ecbbf-ac5c-4924-a396-549da1855e65/nova-metadata-metadata/0.log" Dec 01 09:40:56 crc kubenswrapper[4873]: I1201 09:40:56.451797 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_43a97c60-b43e-4896-9592-d41863b1c203/openstackclient/0.log" Dec 01 09:40:56 crc kubenswrapper[4873]: I1201 09:40:56.604877 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-7xdqf_93fb09e4-51b3-44f7-afab-c49c374c7f56/openstack-network-exporter/0.log" Dec 01 09:40:56 crc kubenswrapper[4873]: I1201 09:40:56.766346 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-xd66n_243fec55-bb19-4807-8b15-69584a922cf6/ovsdb-server-init/0.log" Dec 01 09:40:56 crc kubenswrapper[4873]: I1201 09:40:56.970692 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-xd66n_243fec55-bb19-4807-8b15-69584a922cf6/ovs-vswitchd/0.log" Dec 01 09:40:57 crc kubenswrapper[4873]: I1201 09:40:57.001671 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-xd66n_243fec55-bb19-4807-8b15-69584a922cf6/ovsdb-server-init/0.log" Dec 01 09:40:57 crc kubenswrapper[4873]: I1201 09:40:57.006871 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-xd66n_243fec55-bb19-4807-8b15-69584a922cf6/ovsdb-server/0.log" Dec 01 09:40:57 crc kubenswrapper[4873]: I1201 09:40:57.223345 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-w8v6j_b79364c5-00b2-4ab8-b2f0-aff17ed902b4/ovn-controller/0.log" Dec 01 09:40:57 crc kubenswrapper[4873]: I1201 09:40:57.295399 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-q7cl2_0b014b12-82de-4f51-84bc-c8d41241f672/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:40:57 crc kubenswrapper[4873]: I1201 09:40:57.476834 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_a5cf24b4-dd05-45a4-bf8f-5895c30ada04/openstack-network-exporter/0.log" Dec 01 09:40:57 crc kubenswrapper[4873]: I1201 09:40:57.565758 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_a5cf24b4-dd05-45a4-bf8f-5895c30ada04/ovn-northd/0.log" Dec 01 09:40:57 crc kubenswrapper[4873]: I1201 09:40:57.761458 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_21c27fff-80e6-4a74-aad2-5e772b811535/openstack-network-exporter/0.log" Dec 01 09:40:57 crc kubenswrapper[4873]: I1201 09:40:57.817584 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_21c27fff-80e6-4a74-aad2-5e772b811535/ovsdbserver-nb/0.log" Dec 01 09:40:58 crc kubenswrapper[4873]: I1201 09:40:58.010524 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_dbe0e07d-50a3-47c2-831d-69f24c0e838a/openstack-network-exporter/0.log" Dec 01 09:40:58 crc kubenswrapper[4873]: I1201 09:40:58.042485 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_dbe0e07d-50a3-47c2-831d-69f24c0e838a/ovsdbserver-sb/0.log" Dec 01 09:40:58 crc kubenswrapper[4873]: I1201 09:40:58.198564 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7d689bd78d-s9rtj_6615f806-2d35-4500-a746-3fab7fbfde88/placement-api/0.log" Dec 01 09:40:58 crc kubenswrapper[4873]: I1201 09:40:58.384622 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7d689bd78d-s9rtj_6615f806-2d35-4500-a746-3fab7fbfde88/placement-log/0.log" Dec 01 09:40:58 crc kubenswrapper[4873]: I1201 09:40:58.444443 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3baa8150-922b-4de7-ae44-24b9b5584212/setup-container/0.log" Dec 01 09:40:58 crc kubenswrapper[4873]: I1201 09:40:58.690348 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3baa8150-922b-4de7-ae44-24b9b5584212/setup-container/0.log" Dec 01 09:40:58 crc kubenswrapper[4873]: I1201 09:40:58.706437 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3baa8150-922b-4de7-ae44-24b9b5584212/rabbitmq/0.log" Dec 01 09:40:58 crc kubenswrapper[4873]: I1201 09:40:58.724552 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_84aac4b8-9ed2-40d2-81f1-eba3af6e46b0/setup-container/0.log" Dec 01 09:40:59 crc kubenswrapper[4873]: I1201 09:40:59.023316 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_84aac4b8-9ed2-40d2-81f1-eba3af6e46b0/setup-container/0.log" Dec 01 09:40:59 crc kubenswrapper[4873]: I1201 09:40:59.104230 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_84aac4b8-9ed2-40d2-81f1-eba3af6e46b0/rabbitmq/0.log" Dec 01 09:40:59 crc kubenswrapper[4873]: I1201 09:40:59.134293 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-9bdgf_e301f3ff-9983-4440-ad43-96ef975437f7/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:40:59 crc kubenswrapper[4873]: I1201 09:40:59.404759 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-gg2dk_74115943-1cf4-48ce-9f8c-41de87f8fe52/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:40:59 crc kubenswrapper[4873]: I1201 09:40:59.429753 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-g5ghr_47baf114-fcb4-46ba-ac64-53f969ebcbbc/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:40:59 crc kubenswrapper[4873]: I1201 09:40:59.649231 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-9k6m8_2ea95238-7e66-46b0-9c20-b6088f9d6737/ssh-known-hosts-edpm-deployment/0.log" Dec 01 09:40:59 crc kubenswrapper[4873]: I1201 09:40:59.773975 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_cd15a962-a485-48ad-b0af-d7a3ec0e4ab1/tempest-tests-tempest-tests-runner/0.log" Dec 01 09:41:00 crc kubenswrapper[4873]: I1201 09:41:00.381591 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_f210543d-a2a6-4746-b279-04fe37ee1b76/test-operator-logs-container/0.log" Dec 01 09:41:00 crc kubenswrapper[4873]: I1201 09:41:00.521244 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-qt48t_933df0ea-eae8-49a3-affc-2f421e3ae777/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:41:06 crc kubenswrapper[4873]: I1201 09:41:06.430166 4873 scope.go:117] "RemoveContainer" containerID="c101914cc0d7462afe57653cc6d66f5335bbd707b5b5acec3231122718624d93" Dec 01 09:41:06 crc kubenswrapper[4873]: E1201 09:41:06.431213 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:41:17 crc kubenswrapper[4873]: I1201 09:41:17.430471 4873 scope.go:117] "RemoveContainer" containerID="c101914cc0d7462afe57653cc6d66f5335bbd707b5b5acec3231122718624d93" Dec 01 09:41:17 crc kubenswrapper[4873]: E1201 09:41:17.431394 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:41:20 crc kubenswrapper[4873]: I1201 09:41:20.834538 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_2e6c09a7-97f9-4048-b5a9-b3df685eacbc/memcached/0.log" Dec 01 09:41:29 crc kubenswrapper[4873]: I1201 09:41:29.785844 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-5rq8z_51253ba6-e0b5-44ac-8c18-be17b4d13024/kube-rbac-proxy/0.log" Dec 01 09:41:29 crc kubenswrapper[4873]: I1201 09:41:29.889572 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-5rq8z_51253ba6-e0b5-44ac-8c18-be17b4d13024/manager/0.log" Dec 01 09:41:30 crc kubenswrapper[4873]: I1201 09:41:30.097104 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-gqm5w_671ad3cc-9716-42de-b5ae-fb69847d3bd7/kube-rbac-proxy/0.log" Dec 01 09:41:30 crc kubenswrapper[4873]: I1201 09:41:30.201367 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-gqm5w_671ad3cc-9716-42de-b5ae-fb69847d3bd7/manager/0.log" Dec 01 09:41:30 crc kubenswrapper[4873]: I1201 09:41:30.252075 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d92c7d608fded6ff7325ca25db46cfa3dc2f9a2773cb154702b56fef16ks4c9_08dfa43f-80d7-4785-a0ce-c8b6d8d95351/util/0.log" Dec 01 09:41:30 crc kubenswrapper[4873]: I1201 09:41:30.507546 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d92c7d608fded6ff7325ca25db46cfa3dc2f9a2773cb154702b56fef16ks4c9_08dfa43f-80d7-4785-a0ce-c8b6d8d95351/pull/0.log" Dec 01 09:41:30 crc kubenswrapper[4873]: I1201 09:41:30.516430 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d92c7d608fded6ff7325ca25db46cfa3dc2f9a2773cb154702b56fef16ks4c9_08dfa43f-80d7-4785-a0ce-c8b6d8d95351/util/0.log" Dec 01 09:41:30 crc kubenswrapper[4873]: I1201 09:41:30.530854 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d92c7d608fded6ff7325ca25db46cfa3dc2f9a2773cb154702b56fef16ks4c9_08dfa43f-80d7-4785-a0ce-c8b6d8d95351/pull/0.log" Dec 01 09:41:30 crc kubenswrapper[4873]: I1201 09:41:30.697100 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d92c7d608fded6ff7325ca25db46cfa3dc2f9a2773cb154702b56fef16ks4c9_08dfa43f-80d7-4785-a0ce-c8b6d8d95351/pull/0.log" Dec 01 09:41:30 crc kubenswrapper[4873]: I1201 09:41:30.702633 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d92c7d608fded6ff7325ca25db46cfa3dc2f9a2773cb154702b56fef16ks4c9_08dfa43f-80d7-4785-a0ce-c8b6d8d95351/util/0.log" Dec 01 09:41:30 crc kubenswrapper[4873]: I1201 09:41:30.742376 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d92c7d608fded6ff7325ca25db46cfa3dc2f9a2773cb154702b56fef16ks4c9_08dfa43f-80d7-4785-a0ce-c8b6d8d95351/extract/0.log" Dec 01 09:41:30 crc kubenswrapper[4873]: I1201 09:41:30.909211 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-6r8hb_723f446d-21bb-432a-b2af-93219593819d/manager/0.log" Dec 01 09:41:30 crc kubenswrapper[4873]: I1201 09:41:30.919121 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-6r8hb_723f446d-21bb-432a-b2af-93219593819d/kube-rbac-proxy/0.log" Dec 01 09:41:31 crc kubenswrapper[4873]: I1201 09:41:31.011703 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5cd474b4ff-wcmv4_f778fd6c-4c8a-4067-b3f2-cb7d98a50bf7/kube-rbac-proxy/0.log" Dec 01 09:41:31 crc kubenswrapper[4873]: I1201 09:41:31.179913 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5cd474b4ff-wcmv4_f778fd6c-4c8a-4067-b3f2-cb7d98a50bf7/manager/0.log" Dec 01 09:41:31 crc kubenswrapper[4873]: I1201 09:41:31.198271 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-d7c6w_967215e8-7a18-4507-96c1-4c79c7e1d51a/kube-rbac-proxy/0.log" Dec 01 09:41:31 crc kubenswrapper[4873]: I1201 09:41:31.337870 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-d7c6w_967215e8-7a18-4507-96c1-4c79c7e1d51a/manager/0.log" Dec 01 09:41:31 crc kubenswrapper[4873]: I1201 09:41:31.434327 4873 scope.go:117] "RemoveContainer" containerID="c101914cc0d7462afe57653cc6d66f5335bbd707b5b5acec3231122718624d93" Dec 01 09:41:31 crc kubenswrapper[4873]: E1201 09:41:31.435463 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:41:31 crc kubenswrapper[4873]: I1201 09:41:31.490368 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-45whz_94e7345b-b1ac-46e2-be25-9d64d3d33523/kube-rbac-proxy/0.log" Dec 01 09:41:31 crc kubenswrapper[4873]: I1201 09:41:31.526051 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-45whz_94e7345b-b1ac-46e2-be25-9d64d3d33523/manager/0.log" Dec 01 09:41:31 crc kubenswrapper[4873]: I1201 09:41:31.683883 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-tmdj7_4dcc939e-b548-4fb6-814a-30e2aaa8a94a/kube-rbac-proxy/0.log" Dec 01 09:41:31 crc kubenswrapper[4873]: I1201 09:41:31.784800 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-7wmqg_eb53d542-63e9-487d-9d06-237c4b2b9252/kube-rbac-proxy/0.log" Dec 01 09:41:31 crc kubenswrapper[4873]: I1201 09:41:31.933215 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-tmdj7_4dcc939e-b548-4fb6-814a-30e2aaa8a94a/manager/0.log" Dec 01 09:41:31 crc kubenswrapper[4873]: I1201 09:41:31.935180 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-7wmqg_eb53d542-63e9-487d-9d06-237c4b2b9252/manager/0.log" Dec 01 09:41:32 crc kubenswrapper[4873]: I1201 09:41:32.011755 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-546d4bdf48-pfphw_035095bd-1b65-4895-95a1-59feee524920/kube-rbac-proxy/0.log" Dec 01 09:41:32 crc kubenswrapper[4873]: I1201 09:41:32.137922 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6546668bfd-j2vnt_b45909ce-2ea4-4cf2-9351-c8839e44d734/kube-rbac-proxy/0.log" Dec 01 09:41:32 crc kubenswrapper[4873]: I1201 09:41:32.214625 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-546d4bdf48-pfphw_035095bd-1b65-4895-95a1-59feee524920/manager/0.log" Dec 01 09:41:32 crc kubenswrapper[4873]: I1201 09:41:32.304598 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6546668bfd-j2vnt_b45909ce-2ea4-4cf2-9351-c8839e44d734/manager/0.log" Dec 01 09:41:32 crc kubenswrapper[4873]: I1201 09:41:32.407219 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-hnzzf_3eff7ce2-9ff7-413c-b472-9e114e7130ca/kube-rbac-proxy/0.log" Dec 01 09:41:32 crc kubenswrapper[4873]: I1201 09:41:32.439424 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-hnzzf_3eff7ce2-9ff7-413c-b472-9e114e7130ca/manager/0.log" Dec 01 09:41:32 crc kubenswrapper[4873]: I1201 09:41:32.687115 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-sjw48_5ec1188a-c7f9-4cc8-8ea9-c1f1977041a5/kube-rbac-proxy/0.log" Dec 01 09:41:32 crc kubenswrapper[4873]: I1201 09:41:32.702734 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-sjw48_5ec1188a-c7f9-4cc8-8ea9-c1f1977041a5/manager/0.log" Dec 01 09:41:32 crc kubenswrapper[4873]: I1201 09:41:32.823065 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-x9kc4_450a3417-0074-4223-b1e3-aa1b854320fe/kube-rbac-proxy/0.log" Dec 01 09:41:32 crc kubenswrapper[4873]: I1201 09:41:32.923978 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-n79q9_7e3a2af8-3381-46e2-8c23-41aab8fd1a5e/kube-rbac-proxy/0.log" Dec 01 09:41:33 crc kubenswrapper[4873]: I1201 09:41:33.014335 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-x9kc4_450a3417-0074-4223-b1e3-aa1b854320fe/manager/0.log" Dec 01 09:41:33 crc kubenswrapper[4873]: I1201 09:41:33.060363 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-n79q9_7e3a2af8-3381-46e2-8c23-41aab8fd1a5e/manager/0.log" Dec 01 09:41:33 crc kubenswrapper[4873]: I1201 09:41:33.158691 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4vf8qt_5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5/kube-rbac-proxy/0.log" Dec 01 09:41:33 crc kubenswrapper[4873]: I1201 09:41:33.252687 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4vf8qt_5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5/manager/0.log" Dec 01 09:41:33 crc kubenswrapper[4873]: I1201 09:41:33.679405 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-mqm6p_b1309d4c-45e2-444c-afc5-56aa75b9abf5/registry-server/0.log" Dec 01 09:41:33 crc kubenswrapper[4873]: I1201 09:41:33.725659 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-7f57cdcc9-jttc7_c25dac67-107f-43e2-a63e-5843ff31abc8/operator/0.log" Dec 01 09:41:33 crc kubenswrapper[4873]: I1201 09:41:33.934145 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-49bz9_879e02ea-306e-4e7b-9012-cb1cb6bdee00/kube-rbac-proxy/0.log" Dec 01 09:41:33 crc kubenswrapper[4873]: I1201 09:41:33.993259 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-49bz9_879e02ea-306e-4e7b-9012-cb1cb6bdee00/manager/0.log" Dec 01 09:41:33 crc kubenswrapper[4873]: I1201 09:41:33.995813 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-v7bvz_3e0ad069-d971-4fd0-93df-b6a8794afe00/kube-rbac-proxy/0.log" Dec 01 09:41:34 crc kubenswrapper[4873]: I1201 09:41:34.237212 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-cb9qn_664e20ef-b15c-4903-b72d-e18c7077e888/operator/0.log" Dec 01 09:41:34 crc kubenswrapper[4873]: I1201 09:41:34.238386 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-v7bvz_3e0ad069-d971-4fd0-93df-b6a8794afe00/manager/0.log" Dec 01 09:41:34 crc kubenswrapper[4873]: I1201 09:41:34.428270 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-6dxcw_f7a7bcca-9403-4e2d-bc28-53d4eb5ae252/kube-rbac-proxy/0.log" Dec 01 09:41:34 crc kubenswrapper[4873]: I1201 09:41:34.537529 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-6dxcw_f7a7bcca-9403-4e2d-bc28-53d4eb5ae252/manager/0.log" Dec 01 09:41:34 crc kubenswrapper[4873]: I1201 09:41:34.558727 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-bhrpn_225323c8-ed2f-4573-920d-f43f3e4561de/kube-rbac-proxy/0.log" Dec 01 09:41:34 crc kubenswrapper[4873]: I1201 09:41:34.765887 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-bhrpn_225323c8-ed2f-4573-920d-f43f3e4561de/manager/0.log" Dec 01 09:41:34 crc kubenswrapper[4873]: I1201 09:41:34.803503 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7f4d6cf897-fddkc_223a64ee-92cb-4b29-91bf-ffa7ed7d64ce/manager/0.log" Dec 01 09:41:34 crc kubenswrapper[4873]: I1201 09:41:34.830533 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-mvc26_34e3726b-c9f5-4a86-9e87-e179f8dae739/manager/0.log" Dec 01 09:41:34 crc kubenswrapper[4873]: I1201 09:41:34.834422 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-mvc26_34e3726b-c9f5-4a86-9e87-e179f8dae739/kube-rbac-proxy/0.log" Dec 01 09:41:34 crc kubenswrapper[4873]: I1201 09:41:34.993545 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-dx4jm_1cd3aa85-6333-4cfd-aacf-e51169ef4b42/manager/0.log" Dec 01 09:41:34 crc kubenswrapper[4873]: I1201 09:41:34.998765 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-dx4jm_1cd3aa85-6333-4cfd-aacf-e51169ef4b42/kube-rbac-proxy/0.log" Dec 01 09:41:43 crc kubenswrapper[4873]: I1201 09:41:43.431360 4873 scope.go:117] "RemoveContainer" containerID="c101914cc0d7462afe57653cc6d66f5335bbd707b5b5acec3231122718624d93" Dec 01 09:41:43 crc kubenswrapper[4873]: E1201 09:41:43.432455 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:41:54 crc kubenswrapper[4873]: I1201 09:41:54.089980 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-xg89d_f4709124-2f04-4d3c-abb7-30a5d0925b6c/control-plane-machine-set-operator/0.log" Dec 01 09:41:54 crc kubenswrapper[4873]: I1201 09:41:54.255633 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-52pjn_fb8fcfca-0933-4a74-85de-859fda0153d6/kube-rbac-proxy/0.log" Dec 01 09:41:54 crc kubenswrapper[4873]: I1201 09:41:54.294131 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-52pjn_fb8fcfca-0933-4a74-85de-859fda0153d6/machine-api-operator/0.log" Dec 01 09:41:56 crc kubenswrapper[4873]: I1201 09:41:56.430060 4873 scope.go:117] "RemoveContainer" containerID="c101914cc0d7462afe57653cc6d66f5335bbd707b5b5acec3231122718624d93" Dec 01 09:41:56 crc kubenswrapper[4873]: E1201 09:41:56.430707 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:42:06 crc kubenswrapper[4873]: I1201 09:42:06.974047 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-blv72_c52cc19c-1588-49cd-a54b-072c33505a5a/cert-manager-controller/0.log" Dec 01 09:42:07 crc kubenswrapper[4873]: I1201 09:42:07.112182 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-47tnm_3876e535-8b83-41bf-8b75-2b21a25e1ff4/cert-manager-cainjector/0.log" Dec 01 09:42:07 crc kubenswrapper[4873]: I1201 09:42:07.164781 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-tj5cg_eb0ae03d-2543-42c9-9c93-628cab31efe2/cert-manager-webhook/0.log" Dec 01 09:42:11 crc kubenswrapper[4873]: I1201 09:42:11.430701 4873 scope.go:117] "RemoveContainer" containerID="c101914cc0d7462afe57653cc6d66f5335bbd707b5b5acec3231122718624d93" Dec 01 09:42:11 crc kubenswrapper[4873]: E1201 09:42:11.431611 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:42:19 crc kubenswrapper[4873]: I1201 09:42:19.802396 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-rtjft_244acac6-afd2-4d7b-b1a0-085a24cbf1c8/nmstate-console-plugin/0.log" Dec 01 09:42:20 crc kubenswrapper[4873]: I1201 09:42:20.073838 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-m7d79_1c32a501-0496-45b2-8e04-984bccb3c03d/nmstate-metrics/0.log" Dec 01 09:42:20 crc kubenswrapper[4873]: I1201 09:42:20.087769 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-x8slw_5ca9fe36-a6a0-402c-a744-4125d0cd37c6/nmstate-handler/0.log" Dec 01 09:42:20 crc kubenswrapper[4873]: I1201 09:42:20.130009 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-m7d79_1c32a501-0496-45b2-8e04-984bccb3c03d/kube-rbac-proxy/0.log" Dec 01 09:42:20 crc kubenswrapper[4873]: I1201 09:42:20.339992 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-fbds9_a61c7c72-a083-41c2-b0a4-707d26b095c6/nmstate-operator/0.log" Dec 01 09:42:20 crc kubenswrapper[4873]: I1201 09:42:20.425173 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-79fqx_6050180f-b828-43a2-911c-4a44354d4bf6/nmstate-webhook/0.log" Dec 01 09:42:22 crc kubenswrapper[4873]: I1201 09:42:22.435953 4873 scope.go:117] "RemoveContainer" containerID="c101914cc0d7462afe57653cc6d66f5335bbd707b5b5acec3231122718624d93" Dec 01 09:42:22 crc kubenswrapper[4873]: E1201 09:42:22.436842 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:42:34 crc kubenswrapper[4873]: I1201 09:42:34.441081 4873 scope.go:117] "RemoveContainer" containerID="c101914cc0d7462afe57653cc6d66f5335bbd707b5b5acec3231122718624d93" Dec 01 09:42:34 crc kubenswrapper[4873]: E1201 09:42:34.441978 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:42:35 crc kubenswrapper[4873]: I1201 09:42:35.565702 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-6wb84_0bd118a2-14e2-438a-bea0-2cd777f71e5c/controller/0.log" Dec 01 09:42:35 crc kubenswrapper[4873]: I1201 09:42:35.570554 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-6wb84_0bd118a2-14e2-438a-bea0-2cd777f71e5c/kube-rbac-proxy/0.log" Dec 01 09:42:35 crc kubenswrapper[4873]: I1201 09:42:35.742322 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v584n_065b9940-14f0-45e2-9d92-3e7173fce9b2/cp-frr-files/0.log" Dec 01 09:42:35 crc kubenswrapper[4873]: I1201 09:42:35.938644 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v584n_065b9940-14f0-45e2-9d92-3e7173fce9b2/cp-frr-files/0.log" Dec 01 09:42:35 crc kubenswrapper[4873]: I1201 09:42:35.975862 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v584n_065b9940-14f0-45e2-9d92-3e7173fce9b2/cp-reloader/0.log" Dec 01 09:42:35 crc kubenswrapper[4873]: I1201 09:42:35.977238 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v584n_065b9940-14f0-45e2-9d92-3e7173fce9b2/cp-metrics/0.log" Dec 01 09:42:35 crc kubenswrapper[4873]: I1201 09:42:35.979400 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v584n_065b9940-14f0-45e2-9d92-3e7173fce9b2/cp-reloader/0.log" Dec 01 09:42:36 crc kubenswrapper[4873]: I1201 09:42:36.179878 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v584n_065b9940-14f0-45e2-9d92-3e7173fce9b2/cp-metrics/0.log" Dec 01 09:42:36 crc kubenswrapper[4873]: I1201 09:42:36.179890 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v584n_065b9940-14f0-45e2-9d92-3e7173fce9b2/cp-metrics/0.log" Dec 01 09:42:36 crc kubenswrapper[4873]: I1201 09:42:36.226775 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v584n_065b9940-14f0-45e2-9d92-3e7173fce9b2/cp-reloader/0.log" Dec 01 09:42:36 crc kubenswrapper[4873]: I1201 09:42:36.226882 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v584n_065b9940-14f0-45e2-9d92-3e7173fce9b2/cp-frr-files/0.log" Dec 01 09:42:36 crc kubenswrapper[4873]: I1201 09:42:36.421295 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v584n_065b9940-14f0-45e2-9d92-3e7173fce9b2/cp-reloader/0.log" Dec 01 09:42:36 crc kubenswrapper[4873]: I1201 09:42:36.465809 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v584n_065b9940-14f0-45e2-9d92-3e7173fce9b2/cp-metrics/0.log" Dec 01 09:42:36 crc kubenswrapper[4873]: I1201 09:42:36.465811 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v584n_065b9940-14f0-45e2-9d92-3e7173fce9b2/cp-frr-files/0.log" Dec 01 09:42:36 crc kubenswrapper[4873]: I1201 09:42:36.480837 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v584n_065b9940-14f0-45e2-9d92-3e7173fce9b2/controller/0.log" Dec 01 09:42:36 crc kubenswrapper[4873]: I1201 09:42:36.696096 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v584n_065b9940-14f0-45e2-9d92-3e7173fce9b2/frr-metrics/0.log" Dec 01 09:42:36 crc kubenswrapper[4873]: I1201 09:42:36.754902 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v584n_065b9940-14f0-45e2-9d92-3e7173fce9b2/kube-rbac-proxy-frr/0.log" Dec 01 09:42:36 crc kubenswrapper[4873]: I1201 09:42:36.772495 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v584n_065b9940-14f0-45e2-9d92-3e7173fce9b2/kube-rbac-proxy/0.log" Dec 01 09:42:37 crc kubenswrapper[4873]: I1201 09:42:37.160131 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v584n_065b9940-14f0-45e2-9d92-3e7173fce9b2/reloader/0.log" Dec 01 09:42:37 crc kubenswrapper[4873]: I1201 09:42:37.232730 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-ffvzp_31ad102b-37f9-491d-b017-20c6a3bc4973/frr-k8s-webhook-server/0.log" Dec 01 09:42:37 crc kubenswrapper[4873]: I1201 09:42:37.497522 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-586dc797b5-msvs2_5b1a0fb8-99da-4202-8c8e-6f5fbcec1908/manager/0.log" Dec 01 09:42:37 crc kubenswrapper[4873]: I1201 09:42:37.827833 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-cbkfq_8f181c2f-7d70-4dd9-bcab-ae125dfa4037/kube-rbac-proxy/0.log" Dec 01 09:42:37 crc kubenswrapper[4873]: I1201 09:42:37.869858 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-767c58684d-z249t_c9747d5c-eed2-4d46-8cbf-0a01c26624da/webhook-server/0.log" Dec 01 09:42:38 crc kubenswrapper[4873]: I1201 09:42:38.286679 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v584n_065b9940-14f0-45e2-9d92-3e7173fce9b2/frr/0.log" Dec 01 09:42:38 crc kubenswrapper[4873]: I1201 09:42:38.470116 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-cbkfq_8f181c2f-7d70-4dd9-bcab-ae125dfa4037/speaker/0.log" Dec 01 09:42:48 crc kubenswrapper[4873]: I1201 09:42:48.432976 4873 scope.go:117] "RemoveContainer" containerID="c101914cc0d7462afe57653cc6d66f5335bbd707b5b5acec3231122718624d93" Dec 01 09:42:48 crc kubenswrapper[4873]: E1201 09:42:48.434548 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:42:51 crc kubenswrapper[4873]: I1201 09:42:51.197729 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhgjjb_b77a8d6f-5a83-494e-a366-300ffde2afb4/util/0.log" Dec 01 09:42:51 crc kubenswrapper[4873]: I1201 09:42:51.438173 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhgjjb_b77a8d6f-5a83-494e-a366-300ffde2afb4/pull/0.log" Dec 01 09:42:51 crc kubenswrapper[4873]: I1201 09:42:51.444059 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhgjjb_b77a8d6f-5a83-494e-a366-300ffde2afb4/util/0.log" Dec 01 09:42:51 crc kubenswrapper[4873]: I1201 09:42:51.471262 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhgjjb_b77a8d6f-5a83-494e-a366-300ffde2afb4/pull/0.log" Dec 01 09:42:51 crc kubenswrapper[4873]: I1201 09:42:51.627515 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhgjjb_b77a8d6f-5a83-494e-a366-300ffde2afb4/util/0.log" Dec 01 09:42:51 crc kubenswrapper[4873]: I1201 09:42:51.660943 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhgjjb_b77a8d6f-5a83-494e-a366-300ffde2afb4/pull/0.log" Dec 01 09:42:51 crc kubenswrapper[4873]: I1201 09:42:51.670417 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhgjjb_b77a8d6f-5a83-494e-a366-300ffde2afb4/extract/0.log" Dec 01 09:42:51 crc kubenswrapper[4873]: I1201 09:42:51.833130 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8359xjq_8d9127f2-7870-4647-8655-0c8fff4df500/util/0.log" Dec 01 09:42:52 crc kubenswrapper[4873]: I1201 09:42:52.006158 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8359xjq_8d9127f2-7870-4647-8655-0c8fff4df500/pull/0.log" Dec 01 09:42:52 crc kubenswrapper[4873]: I1201 09:42:52.047756 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8359xjq_8d9127f2-7870-4647-8655-0c8fff4df500/util/0.log" Dec 01 09:42:52 crc kubenswrapper[4873]: I1201 09:42:52.062097 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-9a52-account-create-update-vtg5d"] Dec 01 09:42:52 crc kubenswrapper[4873]: I1201 09:42:52.076632 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-9a52-account-create-update-vtg5d"] Dec 01 09:42:52 crc kubenswrapper[4873]: I1201 09:42:52.087520 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-create-v7b4d"] Dec 01 09:42:52 crc kubenswrapper[4873]: I1201 09:42:52.095051 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-create-v7b4d"] Dec 01 09:42:52 crc kubenswrapper[4873]: I1201 09:42:52.098642 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8359xjq_8d9127f2-7870-4647-8655-0c8fff4df500/pull/0.log" Dec 01 09:42:52 crc kubenswrapper[4873]: I1201 09:42:52.273551 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8359xjq_8d9127f2-7870-4647-8655-0c8fff4df500/util/0.log" Dec 01 09:42:52 crc kubenswrapper[4873]: I1201 09:42:52.279346 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8359xjq_8d9127f2-7870-4647-8655-0c8fff4df500/pull/0.log" Dec 01 09:42:52 crc kubenswrapper[4873]: I1201 09:42:52.301182 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8359xjq_8d9127f2-7870-4647-8655-0c8fff4df500/extract/0.log" Dec 01 09:42:52 crc kubenswrapper[4873]: I1201 09:42:52.447621 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="483831f0-8b26-4845-b893-50bb7fb6ddb4" path="/var/lib/kubelet/pods/483831f0-8b26-4845-b893-50bb7fb6ddb4/volumes" Dec 01 09:42:52 crc kubenswrapper[4873]: I1201 09:42:52.448549 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8a98544-560c-487a-9a71-59b53a882c9b" path="/var/lib/kubelet/pods/a8a98544-560c-487a-9a71-59b53a882c9b/volumes" Dec 01 09:42:52 crc kubenswrapper[4873]: I1201 09:42:52.475407 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pdgt5_96e1f659-b44b-4573-ba28-9605403487e5/extract-utilities/0.log" Dec 01 09:42:52 crc kubenswrapper[4873]: I1201 09:42:52.696700 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pdgt5_96e1f659-b44b-4573-ba28-9605403487e5/extract-content/0.log" Dec 01 09:42:52 crc kubenswrapper[4873]: I1201 09:42:52.711973 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pdgt5_96e1f659-b44b-4573-ba28-9605403487e5/extract-content/0.log" Dec 01 09:42:52 crc kubenswrapper[4873]: I1201 09:42:52.716408 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pdgt5_96e1f659-b44b-4573-ba28-9605403487e5/extract-utilities/0.log" Dec 01 09:42:52 crc kubenswrapper[4873]: I1201 09:42:52.914707 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pdgt5_96e1f659-b44b-4573-ba28-9605403487e5/extract-utilities/0.log" Dec 01 09:42:52 crc kubenswrapper[4873]: I1201 09:42:52.965840 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pdgt5_96e1f659-b44b-4573-ba28-9605403487e5/extract-content/0.log" Dec 01 09:42:53 crc kubenswrapper[4873]: I1201 09:42:53.222448 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7zq5j_3adbfc1d-7e40-4bfc-ab02-bcec5b0a5c38/extract-utilities/0.log" Dec 01 09:42:53 crc kubenswrapper[4873]: I1201 09:42:53.406391 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7zq5j_3adbfc1d-7e40-4bfc-ab02-bcec5b0a5c38/extract-utilities/0.log" Dec 01 09:42:53 crc kubenswrapper[4873]: I1201 09:42:53.444381 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7zq5j_3adbfc1d-7e40-4bfc-ab02-bcec5b0a5c38/extract-content/0.log" Dec 01 09:42:53 crc kubenswrapper[4873]: I1201 09:42:53.490479 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pdgt5_96e1f659-b44b-4573-ba28-9605403487e5/registry-server/0.log" Dec 01 09:42:53 crc kubenswrapper[4873]: I1201 09:42:53.534215 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7zq5j_3adbfc1d-7e40-4bfc-ab02-bcec5b0a5c38/extract-content/0.log" Dec 01 09:42:53 crc kubenswrapper[4873]: I1201 09:42:53.685957 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7zq5j_3adbfc1d-7e40-4bfc-ab02-bcec5b0a5c38/extract-utilities/0.log" Dec 01 09:42:53 crc kubenswrapper[4873]: I1201 09:42:53.729144 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7zq5j_3adbfc1d-7e40-4bfc-ab02-bcec5b0a5c38/extract-content/0.log" Dec 01 09:42:53 crc kubenswrapper[4873]: I1201 09:42:53.960083 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-dgdcc_c0e7b6b5-8852-4dec-bbf2-b7247b34a791/marketplace-operator/0.log" Dec 01 09:42:53 crc kubenswrapper[4873]: I1201 09:42:53.965003 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7zq5j_3adbfc1d-7e40-4bfc-ab02-bcec5b0a5c38/registry-server/0.log" Dec 01 09:42:54 crc kubenswrapper[4873]: I1201 09:42:54.111424 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-p98sp_d62ca400-6749-4a3f-94ec-f7a5716584b9/extract-utilities/0.log" Dec 01 09:42:54 crc kubenswrapper[4873]: I1201 09:42:54.325039 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-p98sp_d62ca400-6749-4a3f-94ec-f7a5716584b9/extract-utilities/0.log" Dec 01 09:42:54 crc kubenswrapper[4873]: I1201 09:42:54.325058 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-p98sp_d62ca400-6749-4a3f-94ec-f7a5716584b9/extract-content/0.log" Dec 01 09:42:54 crc kubenswrapper[4873]: I1201 09:42:54.336367 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-p98sp_d62ca400-6749-4a3f-94ec-f7a5716584b9/extract-content/0.log" Dec 01 09:42:54 crc kubenswrapper[4873]: I1201 09:42:54.518121 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-p98sp_d62ca400-6749-4a3f-94ec-f7a5716584b9/extract-content/0.log" Dec 01 09:42:54 crc kubenswrapper[4873]: I1201 09:42:54.523249 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-p98sp_d62ca400-6749-4a3f-94ec-f7a5716584b9/extract-utilities/0.log" Dec 01 09:42:54 crc kubenswrapper[4873]: I1201 09:42:54.677791 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-p98sp_d62ca400-6749-4a3f-94ec-f7a5716584b9/registry-server/0.log" Dec 01 09:42:54 crc kubenswrapper[4873]: I1201 09:42:54.788656 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-v6fx2_3e92483b-2415-4432-9059-3a60badb2e4a/extract-utilities/0.log" Dec 01 09:42:54 crc kubenswrapper[4873]: I1201 09:42:54.934960 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-v6fx2_3e92483b-2415-4432-9059-3a60badb2e4a/extract-utilities/0.log" Dec 01 09:42:54 crc kubenswrapper[4873]: I1201 09:42:54.975470 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-v6fx2_3e92483b-2415-4432-9059-3a60badb2e4a/extract-content/0.log" Dec 01 09:42:54 crc kubenswrapper[4873]: I1201 09:42:54.975521 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-v6fx2_3e92483b-2415-4432-9059-3a60badb2e4a/extract-content/0.log" Dec 01 09:42:55 crc kubenswrapper[4873]: I1201 09:42:55.129250 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-v6fx2_3e92483b-2415-4432-9059-3a60badb2e4a/extract-content/0.log" Dec 01 09:42:55 crc kubenswrapper[4873]: I1201 09:42:55.144086 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-v6fx2_3e92483b-2415-4432-9059-3a60badb2e4a/extract-utilities/0.log" Dec 01 09:42:55 crc kubenswrapper[4873]: I1201 09:42:55.704937 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-v6fx2_3e92483b-2415-4432-9059-3a60badb2e4a/registry-server/0.log" Dec 01 09:43:03 crc kubenswrapper[4873]: I1201 09:43:03.430362 4873 scope.go:117] "RemoveContainer" containerID="c101914cc0d7462afe57653cc6d66f5335bbd707b5b5acec3231122718624d93" Dec 01 09:43:03 crc kubenswrapper[4873]: E1201 09:43:03.431244 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:43:05 crc kubenswrapper[4873]: I1201 09:43:05.805363 4873 scope.go:117] "RemoveContainer" containerID="d351090aab0a1519656e52e7efd5da92255a58fcd1e22b6940ce50fac5ff0e41" Dec 01 09:43:05 crc kubenswrapper[4873]: I1201 09:43:05.833881 4873 scope.go:117] "RemoveContainer" containerID="0e27232c827dbdbca9e3aed00ce658d6ef88d4774e4add85cdbbbdef58bdbc52" Dec 01 09:43:17 crc kubenswrapper[4873]: I1201 09:43:17.429831 4873 scope.go:117] "RemoveContainer" containerID="c101914cc0d7462afe57653cc6d66f5335bbd707b5b5acec3231122718624d93" Dec 01 09:43:17 crc kubenswrapper[4873]: E1201 09:43:17.430783 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:43:30 crc kubenswrapper[4873]: I1201 09:43:30.056132 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-sync-4fq4c"] Dec 01 09:43:30 crc kubenswrapper[4873]: I1201 09:43:30.078075 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-sync-4fq4c"] Dec 01 09:43:30 crc kubenswrapper[4873]: I1201 09:43:30.434402 4873 scope.go:117] "RemoveContainer" containerID="c101914cc0d7462afe57653cc6d66f5335bbd707b5b5acec3231122718624d93" Dec 01 09:43:30 crc kubenswrapper[4873]: E1201 09:43:30.435160 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:43:30 crc kubenswrapper[4873]: I1201 09:43:30.473766 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f07d0052-de5b-4bc4-9b7b-d65cf86edd06" path="/var/lib/kubelet/pods/f07d0052-de5b-4bc4-9b7b-d65cf86edd06/volumes" Dec 01 09:43:30 crc kubenswrapper[4873]: E1201 09:43:30.702080 4873 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.195:60966->38.102.83.195:42675: write tcp 38.102.83.195:60966->38.102.83.195:42675: write: broken pipe Dec 01 09:43:43 crc kubenswrapper[4873]: I1201 09:43:43.430078 4873 scope.go:117] "RemoveContainer" containerID="c101914cc0d7462afe57653cc6d66f5335bbd707b5b5acec3231122718624d93" Dec 01 09:43:43 crc kubenswrapper[4873]: E1201 09:43:43.434259 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:43:57 crc kubenswrapper[4873]: I1201 09:43:57.430507 4873 scope.go:117] "RemoveContainer" containerID="c101914cc0d7462afe57653cc6d66f5335bbd707b5b5acec3231122718624d93" Dec 01 09:43:57 crc kubenswrapper[4873]: E1201 09:43:57.431508 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:44:05 crc kubenswrapper[4873]: I1201 09:44:05.942576 4873 scope.go:117] "RemoveContainer" containerID="839143e7968b30f668bf8c94fe6e5226bd3dc152c8426bde3713d5f48d698806" Dec 01 09:44:09 crc kubenswrapper[4873]: I1201 09:44:09.430718 4873 scope.go:117] "RemoveContainer" containerID="c101914cc0d7462afe57653cc6d66f5335bbd707b5b5acec3231122718624d93" Dec 01 09:44:09 crc kubenswrapper[4873]: E1201 09:44:09.431979 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:44:20 crc kubenswrapper[4873]: I1201 09:44:20.431512 4873 scope.go:117] "RemoveContainer" containerID="c101914cc0d7462afe57653cc6d66f5335bbd707b5b5acec3231122718624d93" Dec 01 09:44:20 crc kubenswrapper[4873]: E1201 09:44:20.432614 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:44:32 crc kubenswrapper[4873]: I1201 09:44:32.436111 4873 scope.go:117] "RemoveContainer" containerID="c101914cc0d7462afe57653cc6d66f5335bbd707b5b5acec3231122718624d93" Dec 01 09:44:32 crc kubenswrapper[4873]: E1201 09:44:32.436926 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:44:43 crc kubenswrapper[4873]: I1201 09:44:43.431263 4873 scope.go:117] "RemoveContainer" containerID="c101914cc0d7462afe57653cc6d66f5335bbd707b5b5acec3231122718624d93" Dec 01 09:44:43 crc kubenswrapper[4873]: E1201 09:44:43.432088 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:44:46 crc kubenswrapper[4873]: I1201 09:44:46.017316 4873 generic.go:334] "Generic (PLEG): container finished" podID="87fffa2f-8ebb-413f-a56d-cfe1d67eea39" containerID="3b94b41bfb72d83a88c0bcb88544f1492a872fadc16ba4d4c3e873ec15d64d3a" exitCode=0 Dec 01 09:44:46 crc kubenswrapper[4873]: I1201 09:44:46.017394 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2kbfg/must-gather-x85kc" event={"ID":"87fffa2f-8ebb-413f-a56d-cfe1d67eea39","Type":"ContainerDied","Data":"3b94b41bfb72d83a88c0bcb88544f1492a872fadc16ba4d4c3e873ec15d64d3a"} Dec 01 09:44:46 crc kubenswrapper[4873]: I1201 09:44:46.018869 4873 scope.go:117] "RemoveContainer" containerID="3b94b41bfb72d83a88c0bcb88544f1492a872fadc16ba4d4c3e873ec15d64d3a" Dec 01 09:44:46 crc kubenswrapper[4873]: I1201 09:44:46.734567 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-2kbfg_must-gather-x85kc_87fffa2f-8ebb-413f-a56d-cfe1d67eea39/gather/0.log" Dec 01 09:44:54 crc kubenswrapper[4873]: I1201 09:44:54.438302 4873 scope.go:117] "RemoveContainer" containerID="c101914cc0d7462afe57653cc6d66f5335bbd707b5b5acec3231122718624d93" Dec 01 09:44:54 crc kubenswrapper[4873]: E1201 09:44:54.440942 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:44:54 crc kubenswrapper[4873]: I1201 09:44:54.747984 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-2kbfg/must-gather-x85kc"] Dec 01 09:44:54 crc kubenswrapper[4873]: I1201 09:44:54.748654 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-2kbfg/must-gather-x85kc" podUID="87fffa2f-8ebb-413f-a56d-cfe1d67eea39" containerName="copy" containerID="cri-o://a6cf1033c5faf41459c9638ee3912415e6056d18d22703fb9f627730fbfc3793" gracePeriod=2 Dec 01 09:44:54 crc kubenswrapper[4873]: I1201 09:44:54.760514 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-2kbfg/must-gather-x85kc"] Dec 01 09:44:55 crc kubenswrapper[4873]: I1201 09:44:55.118093 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-2kbfg_must-gather-x85kc_87fffa2f-8ebb-413f-a56d-cfe1d67eea39/copy/0.log" Dec 01 09:44:55 crc kubenswrapper[4873]: I1201 09:44:55.118891 4873 generic.go:334] "Generic (PLEG): container finished" podID="87fffa2f-8ebb-413f-a56d-cfe1d67eea39" containerID="a6cf1033c5faf41459c9638ee3912415e6056d18d22703fb9f627730fbfc3793" exitCode=143 Dec 01 09:44:55 crc kubenswrapper[4873]: I1201 09:44:55.282408 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-2kbfg_must-gather-x85kc_87fffa2f-8ebb-413f-a56d-cfe1d67eea39/copy/0.log" Dec 01 09:44:55 crc kubenswrapper[4873]: I1201 09:44:55.282925 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2kbfg/must-gather-x85kc" Dec 01 09:44:55 crc kubenswrapper[4873]: I1201 09:44:55.328400 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mttpd\" (UniqueName: \"kubernetes.io/projected/87fffa2f-8ebb-413f-a56d-cfe1d67eea39-kube-api-access-mttpd\") pod \"87fffa2f-8ebb-413f-a56d-cfe1d67eea39\" (UID: \"87fffa2f-8ebb-413f-a56d-cfe1d67eea39\") " Dec 01 09:44:55 crc kubenswrapper[4873]: I1201 09:44:55.328777 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/87fffa2f-8ebb-413f-a56d-cfe1d67eea39-must-gather-output\") pod \"87fffa2f-8ebb-413f-a56d-cfe1d67eea39\" (UID: \"87fffa2f-8ebb-413f-a56d-cfe1d67eea39\") " Dec 01 09:44:55 crc kubenswrapper[4873]: I1201 09:44:55.336742 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87fffa2f-8ebb-413f-a56d-cfe1d67eea39-kube-api-access-mttpd" (OuterVolumeSpecName: "kube-api-access-mttpd") pod "87fffa2f-8ebb-413f-a56d-cfe1d67eea39" (UID: "87fffa2f-8ebb-413f-a56d-cfe1d67eea39"). InnerVolumeSpecName "kube-api-access-mttpd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:44:55 crc kubenswrapper[4873]: I1201 09:44:55.431797 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mttpd\" (UniqueName: \"kubernetes.io/projected/87fffa2f-8ebb-413f-a56d-cfe1d67eea39-kube-api-access-mttpd\") on node \"crc\" DevicePath \"\"" Dec 01 09:44:55 crc kubenswrapper[4873]: I1201 09:44:55.506715 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87fffa2f-8ebb-413f-a56d-cfe1d67eea39-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "87fffa2f-8ebb-413f-a56d-cfe1d67eea39" (UID: "87fffa2f-8ebb-413f-a56d-cfe1d67eea39"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:44:55 crc kubenswrapper[4873]: I1201 09:44:55.536512 4873 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/87fffa2f-8ebb-413f-a56d-cfe1d67eea39-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 01 09:44:56 crc kubenswrapper[4873]: I1201 09:44:56.133421 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-2kbfg_must-gather-x85kc_87fffa2f-8ebb-413f-a56d-cfe1d67eea39/copy/0.log" Dec 01 09:44:56 crc kubenswrapper[4873]: I1201 09:44:56.135124 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2kbfg/must-gather-x85kc" Dec 01 09:44:56 crc kubenswrapper[4873]: I1201 09:44:56.135127 4873 scope.go:117] "RemoveContainer" containerID="a6cf1033c5faf41459c9638ee3912415e6056d18d22703fb9f627730fbfc3793" Dec 01 09:44:56 crc kubenswrapper[4873]: I1201 09:44:56.163921 4873 scope.go:117] "RemoveContainer" containerID="3b94b41bfb72d83a88c0bcb88544f1492a872fadc16ba4d4c3e873ec15d64d3a" Dec 01 09:44:56 crc kubenswrapper[4873]: I1201 09:44:56.445118 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87fffa2f-8ebb-413f-a56d-cfe1d67eea39" path="/var/lib/kubelet/pods/87fffa2f-8ebb-413f-a56d-cfe1d67eea39/volumes" Dec 01 09:45:00 crc kubenswrapper[4873]: I1201 09:45:00.186272 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409705-j4wfw"] Dec 01 09:45:00 crc kubenswrapper[4873]: E1201 09:45:00.188192 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18da147b-c2cb-4217-887f-d22da4e3e38c" containerName="extract-content" Dec 01 09:45:00 crc kubenswrapper[4873]: I1201 09:45:00.188212 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="18da147b-c2cb-4217-887f-d22da4e3e38c" containerName="extract-content" Dec 01 09:45:00 crc kubenswrapper[4873]: E1201 09:45:00.188227 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87fffa2f-8ebb-413f-a56d-cfe1d67eea39" containerName="gather" Dec 01 09:45:00 crc kubenswrapper[4873]: I1201 09:45:00.188233 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="87fffa2f-8ebb-413f-a56d-cfe1d67eea39" containerName="gather" Dec 01 09:45:00 crc kubenswrapper[4873]: E1201 09:45:00.188247 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1" containerName="extract-utilities" Dec 01 09:45:00 crc kubenswrapper[4873]: I1201 09:45:00.188254 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1" containerName="extract-utilities" Dec 01 09:45:00 crc kubenswrapper[4873]: E1201 09:45:00.188272 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87fffa2f-8ebb-413f-a56d-cfe1d67eea39" containerName="copy" Dec 01 09:45:00 crc kubenswrapper[4873]: I1201 09:45:00.188278 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="87fffa2f-8ebb-413f-a56d-cfe1d67eea39" containerName="copy" Dec 01 09:45:00 crc kubenswrapper[4873]: E1201 09:45:00.188291 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18da147b-c2cb-4217-887f-d22da4e3e38c" containerName="extract-utilities" Dec 01 09:45:00 crc kubenswrapper[4873]: I1201 09:45:00.188296 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="18da147b-c2cb-4217-887f-d22da4e3e38c" containerName="extract-utilities" Dec 01 09:45:00 crc kubenswrapper[4873]: E1201 09:45:00.188305 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18da147b-c2cb-4217-887f-d22da4e3e38c" containerName="registry-server" Dec 01 09:45:00 crc kubenswrapper[4873]: I1201 09:45:00.188313 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="18da147b-c2cb-4217-887f-d22da4e3e38c" containerName="registry-server" Dec 01 09:45:00 crc kubenswrapper[4873]: E1201 09:45:00.188331 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1" containerName="registry-server" Dec 01 09:45:00 crc kubenswrapper[4873]: I1201 09:45:00.188337 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1" containerName="registry-server" Dec 01 09:45:00 crc kubenswrapper[4873]: E1201 09:45:00.188369 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1" containerName="extract-content" Dec 01 09:45:00 crc kubenswrapper[4873]: I1201 09:45:00.188379 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1" containerName="extract-content" Dec 01 09:45:00 crc kubenswrapper[4873]: I1201 09:45:00.188611 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="2dd7bfa7-68a4-43b4-ace6-c7352d63c4a1" containerName="registry-server" Dec 01 09:45:00 crc kubenswrapper[4873]: I1201 09:45:00.188627 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="87fffa2f-8ebb-413f-a56d-cfe1d67eea39" containerName="copy" Dec 01 09:45:00 crc kubenswrapper[4873]: I1201 09:45:00.188642 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="18da147b-c2cb-4217-887f-d22da4e3e38c" containerName="registry-server" Dec 01 09:45:00 crc kubenswrapper[4873]: I1201 09:45:00.188655 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="87fffa2f-8ebb-413f-a56d-cfe1d67eea39" containerName="gather" Dec 01 09:45:00 crc kubenswrapper[4873]: I1201 09:45:00.189685 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409705-j4wfw" Dec 01 09:45:00 crc kubenswrapper[4873]: I1201 09:45:00.201204 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 09:45:00 crc kubenswrapper[4873]: I1201 09:45:00.201596 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 09:45:00 crc kubenswrapper[4873]: I1201 09:45:00.201780 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409705-j4wfw"] Dec 01 09:45:00 crc kubenswrapper[4873]: I1201 09:45:00.301718 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b284c407-6db7-4803-b8d7-74cca8a98f74-config-volume\") pod \"collect-profiles-29409705-j4wfw\" (UID: \"b284c407-6db7-4803-b8d7-74cca8a98f74\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409705-j4wfw" Dec 01 09:45:00 crc kubenswrapper[4873]: I1201 09:45:00.301947 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2fll\" (UniqueName: \"kubernetes.io/projected/b284c407-6db7-4803-b8d7-74cca8a98f74-kube-api-access-w2fll\") pod \"collect-profiles-29409705-j4wfw\" (UID: \"b284c407-6db7-4803-b8d7-74cca8a98f74\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409705-j4wfw" Dec 01 09:45:00 crc kubenswrapper[4873]: I1201 09:45:00.302010 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b284c407-6db7-4803-b8d7-74cca8a98f74-secret-volume\") pod \"collect-profiles-29409705-j4wfw\" (UID: \"b284c407-6db7-4803-b8d7-74cca8a98f74\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409705-j4wfw" Dec 01 09:45:00 crc kubenswrapper[4873]: I1201 09:45:00.404146 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2fll\" (UniqueName: \"kubernetes.io/projected/b284c407-6db7-4803-b8d7-74cca8a98f74-kube-api-access-w2fll\") pod \"collect-profiles-29409705-j4wfw\" (UID: \"b284c407-6db7-4803-b8d7-74cca8a98f74\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409705-j4wfw" Dec 01 09:45:00 crc kubenswrapper[4873]: I1201 09:45:00.404197 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b284c407-6db7-4803-b8d7-74cca8a98f74-secret-volume\") pod \"collect-profiles-29409705-j4wfw\" (UID: \"b284c407-6db7-4803-b8d7-74cca8a98f74\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409705-j4wfw" Dec 01 09:45:00 crc kubenswrapper[4873]: I1201 09:45:00.404293 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b284c407-6db7-4803-b8d7-74cca8a98f74-config-volume\") pod \"collect-profiles-29409705-j4wfw\" (UID: \"b284c407-6db7-4803-b8d7-74cca8a98f74\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409705-j4wfw" Dec 01 09:45:00 crc kubenswrapper[4873]: I1201 09:45:00.405540 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b284c407-6db7-4803-b8d7-74cca8a98f74-config-volume\") pod \"collect-profiles-29409705-j4wfw\" (UID: \"b284c407-6db7-4803-b8d7-74cca8a98f74\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409705-j4wfw" Dec 01 09:45:00 crc kubenswrapper[4873]: I1201 09:45:00.410822 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b284c407-6db7-4803-b8d7-74cca8a98f74-secret-volume\") pod \"collect-profiles-29409705-j4wfw\" (UID: \"b284c407-6db7-4803-b8d7-74cca8a98f74\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409705-j4wfw" Dec 01 09:45:00 crc kubenswrapper[4873]: I1201 09:45:00.427503 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2fll\" (UniqueName: \"kubernetes.io/projected/b284c407-6db7-4803-b8d7-74cca8a98f74-kube-api-access-w2fll\") pod \"collect-profiles-29409705-j4wfw\" (UID: \"b284c407-6db7-4803-b8d7-74cca8a98f74\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409705-j4wfw" Dec 01 09:45:00 crc kubenswrapper[4873]: I1201 09:45:00.520818 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409705-j4wfw" Dec 01 09:45:01 crc kubenswrapper[4873]: I1201 09:45:01.036520 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409705-j4wfw"] Dec 01 09:45:01 crc kubenswrapper[4873]: I1201 09:45:01.205818 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409705-j4wfw" event={"ID":"b284c407-6db7-4803-b8d7-74cca8a98f74","Type":"ContainerStarted","Data":"4b9a745d1732aaa802a2ba481e3738ca8eac7aa7595f93ac465ddad829043fd3"} Dec 01 09:45:02 crc kubenswrapper[4873]: I1201 09:45:02.217399 4873 generic.go:334] "Generic (PLEG): container finished" podID="b284c407-6db7-4803-b8d7-74cca8a98f74" containerID="e949e5fa514b3ef0bb700e245de351e61cfd8313fd3faef465c180791642f886" exitCode=0 Dec 01 09:45:02 crc kubenswrapper[4873]: I1201 09:45:02.217506 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409705-j4wfw" event={"ID":"b284c407-6db7-4803-b8d7-74cca8a98f74","Type":"ContainerDied","Data":"e949e5fa514b3ef0bb700e245de351e61cfd8313fd3faef465c180791642f886"} Dec 01 09:45:03 crc kubenswrapper[4873]: I1201 09:45:03.566784 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409705-j4wfw" Dec 01 09:45:03 crc kubenswrapper[4873]: I1201 09:45:03.685142 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b284c407-6db7-4803-b8d7-74cca8a98f74-secret-volume\") pod \"b284c407-6db7-4803-b8d7-74cca8a98f74\" (UID: \"b284c407-6db7-4803-b8d7-74cca8a98f74\") " Dec 01 09:45:03 crc kubenswrapper[4873]: I1201 09:45:03.685253 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b284c407-6db7-4803-b8d7-74cca8a98f74-config-volume\") pod \"b284c407-6db7-4803-b8d7-74cca8a98f74\" (UID: \"b284c407-6db7-4803-b8d7-74cca8a98f74\") " Dec 01 09:45:03 crc kubenswrapper[4873]: I1201 09:45:03.686266 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b284c407-6db7-4803-b8d7-74cca8a98f74-config-volume" (OuterVolumeSpecName: "config-volume") pod "b284c407-6db7-4803-b8d7-74cca8a98f74" (UID: "b284c407-6db7-4803-b8d7-74cca8a98f74"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:45:03 crc kubenswrapper[4873]: I1201 09:45:03.686347 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w2fll\" (UniqueName: \"kubernetes.io/projected/b284c407-6db7-4803-b8d7-74cca8a98f74-kube-api-access-w2fll\") pod \"b284c407-6db7-4803-b8d7-74cca8a98f74\" (UID: \"b284c407-6db7-4803-b8d7-74cca8a98f74\") " Dec 01 09:45:03 crc kubenswrapper[4873]: I1201 09:45:03.686928 4873 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b284c407-6db7-4803-b8d7-74cca8a98f74-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 09:45:04 crc kubenswrapper[4873]: I1201 09:45:04.368570 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b284c407-6db7-4803-b8d7-74cca8a98f74-kube-api-access-w2fll" (OuterVolumeSpecName: "kube-api-access-w2fll") pod "b284c407-6db7-4803-b8d7-74cca8a98f74" (UID: "b284c407-6db7-4803-b8d7-74cca8a98f74"). InnerVolumeSpecName "kube-api-access-w2fll". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:45:04 crc kubenswrapper[4873]: I1201 09:45:04.370343 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409705-j4wfw" event={"ID":"b284c407-6db7-4803-b8d7-74cca8a98f74","Type":"ContainerDied","Data":"4b9a745d1732aaa802a2ba481e3738ca8eac7aa7595f93ac465ddad829043fd3"} Dec 01 09:45:04 crc kubenswrapper[4873]: I1201 09:45:04.370457 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409705-j4wfw" Dec 01 09:45:04 crc kubenswrapper[4873]: I1201 09:45:04.370552 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4b9a745d1732aaa802a2ba481e3738ca8eac7aa7595f93ac465ddad829043fd3" Dec 01 09:45:04 crc kubenswrapper[4873]: I1201 09:45:04.373007 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b284c407-6db7-4803-b8d7-74cca8a98f74-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b284c407-6db7-4803-b8d7-74cca8a98f74" (UID: "b284c407-6db7-4803-b8d7-74cca8a98f74"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:45:04 crc kubenswrapper[4873]: I1201 09:45:04.466154 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w2fll\" (UniqueName: \"kubernetes.io/projected/b284c407-6db7-4803-b8d7-74cca8a98f74-kube-api-access-w2fll\") on node \"crc\" DevicePath \"\"" Dec 01 09:45:04 crc kubenswrapper[4873]: I1201 09:45:04.466695 4873 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b284c407-6db7-4803-b8d7-74cca8a98f74-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 09:45:04 crc kubenswrapper[4873]: I1201 09:45:04.640132 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409660-hsrhk"] Dec 01 09:45:04 crc kubenswrapper[4873]: I1201 09:45:04.651720 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409660-hsrhk"] Dec 01 09:45:06 crc kubenswrapper[4873]: I1201 09:45:06.444610 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="467ecd19-95ef-4663-b8bd-0a5966bceb01" path="/var/lib/kubelet/pods/467ecd19-95ef-4663-b8bd-0a5966bceb01/volumes" Dec 01 09:45:07 crc kubenswrapper[4873]: I1201 09:45:07.456887 4873 scope.go:117] "RemoveContainer" containerID="c101914cc0d7462afe57653cc6d66f5335bbd707b5b5acec3231122718624d93" Dec 01 09:45:08 crc kubenswrapper[4873]: I1201 09:45:08.477710 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" event={"ID":"fef7b114-0e07-402d-a37b-315c36011f4b","Type":"ContainerStarted","Data":"29490d89d4ce9ae938873b01532387c61bd2469f197fb97051e2685e7e6f8a4e"} Dec 01 09:46:06 crc kubenswrapper[4873]: I1201 09:46:06.093957 4873 scope.go:117] "RemoveContainer" containerID="e7dafb6465a448337fa1b3df8996799cf858e394f863a00d25c0e5d27063e846" Dec 01 09:46:06 crc kubenswrapper[4873]: I1201 09:46:06.127626 4873 scope.go:117] "RemoveContainer" containerID="1d366b6cea0d6f8316e95277ce12c45b342842e883466bf11357e84802083e3f" Dec 01 09:47:31 crc kubenswrapper[4873]: I1201 09:47:31.058609 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:47:31 crc kubenswrapper[4873]: I1201 09:47:31.059352 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:47:38 crc kubenswrapper[4873]: I1201 09:47:38.968295 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mxwtf/must-gather-gbtqv"] Dec 01 09:47:38 crc kubenswrapper[4873]: E1201 09:47:38.969942 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b284c407-6db7-4803-b8d7-74cca8a98f74" containerName="collect-profiles" Dec 01 09:47:38 crc kubenswrapper[4873]: I1201 09:47:38.969963 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="b284c407-6db7-4803-b8d7-74cca8a98f74" containerName="collect-profiles" Dec 01 09:47:38 crc kubenswrapper[4873]: I1201 09:47:38.970257 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="b284c407-6db7-4803-b8d7-74cca8a98f74" containerName="collect-profiles" Dec 01 09:47:38 crc kubenswrapper[4873]: I1201 09:47:38.971706 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mxwtf/must-gather-gbtqv" Dec 01 09:47:38 crc kubenswrapper[4873]: I1201 09:47:38.976481 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-mxwtf"/"kube-root-ca.crt" Dec 01 09:47:38 crc kubenswrapper[4873]: I1201 09:47:38.976798 4873 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-mxwtf"/"default-dockercfg-2j96r" Dec 01 09:47:38 crc kubenswrapper[4873]: I1201 09:47:38.977038 4873 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-mxwtf"/"openshift-service-ca.crt" Dec 01 09:47:38 crc kubenswrapper[4873]: I1201 09:47:38.987421 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-mxwtf/must-gather-gbtqv"] Dec 01 09:47:39 crc kubenswrapper[4873]: I1201 09:47:39.024493 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrxw4\" (UniqueName: \"kubernetes.io/projected/e7f569d0-eb50-4458-9144-b8f4c8947bf1-kube-api-access-nrxw4\") pod \"must-gather-gbtqv\" (UID: \"e7f569d0-eb50-4458-9144-b8f4c8947bf1\") " pod="openshift-must-gather-mxwtf/must-gather-gbtqv" Dec 01 09:47:39 crc kubenswrapper[4873]: I1201 09:47:39.024791 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e7f569d0-eb50-4458-9144-b8f4c8947bf1-must-gather-output\") pod \"must-gather-gbtqv\" (UID: \"e7f569d0-eb50-4458-9144-b8f4c8947bf1\") " pod="openshift-must-gather-mxwtf/must-gather-gbtqv" Dec 01 09:47:39 crc kubenswrapper[4873]: I1201 09:47:39.127156 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e7f569d0-eb50-4458-9144-b8f4c8947bf1-must-gather-output\") pod \"must-gather-gbtqv\" (UID: \"e7f569d0-eb50-4458-9144-b8f4c8947bf1\") " pod="openshift-must-gather-mxwtf/must-gather-gbtqv" Dec 01 09:47:39 crc kubenswrapper[4873]: I1201 09:47:39.127661 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrxw4\" (UniqueName: \"kubernetes.io/projected/e7f569d0-eb50-4458-9144-b8f4c8947bf1-kube-api-access-nrxw4\") pod \"must-gather-gbtqv\" (UID: \"e7f569d0-eb50-4458-9144-b8f4c8947bf1\") " pod="openshift-must-gather-mxwtf/must-gather-gbtqv" Dec 01 09:47:39 crc kubenswrapper[4873]: I1201 09:47:39.129097 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e7f569d0-eb50-4458-9144-b8f4c8947bf1-must-gather-output\") pod \"must-gather-gbtqv\" (UID: \"e7f569d0-eb50-4458-9144-b8f4c8947bf1\") " pod="openshift-must-gather-mxwtf/must-gather-gbtqv" Dec 01 09:47:39 crc kubenswrapper[4873]: I1201 09:47:39.157069 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrxw4\" (UniqueName: \"kubernetes.io/projected/e7f569d0-eb50-4458-9144-b8f4c8947bf1-kube-api-access-nrxw4\") pod \"must-gather-gbtqv\" (UID: \"e7f569d0-eb50-4458-9144-b8f4c8947bf1\") " pod="openshift-must-gather-mxwtf/must-gather-gbtqv" Dec 01 09:47:39 crc kubenswrapper[4873]: I1201 09:47:39.349668 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mxwtf/must-gather-gbtqv" Dec 01 09:47:39 crc kubenswrapper[4873]: I1201 09:47:39.916950 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-mxwtf/must-gather-gbtqv"] Dec 01 09:47:40 crc kubenswrapper[4873]: I1201 09:47:40.216804 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mxwtf/must-gather-gbtqv" event={"ID":"e7f569d0-eb50-4458-9144-b8f4c8947bf1","Type":"ContainerStarted","Data":"c022bfe0bd5cfa09876cfef974dacfb84defea903963a70b944087ab3bb02bdf"} Dec 01 09:47:41 crc kubenswrapper[4873]: I1201 09:47:41.230842 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mxwtf/must-gather-gbtqv" event={"ID":"e7f569d0-eb50-4458-9144-b8f4c8947bf1","Type":"ContainerStarted","Data":"70ca5c7fe0c7a45d30368191d719364c4255c891f9037fa451ac33885a58fb1d"} Dec 01 09:47:41 crc kubenswrapper[4873]: I1201 09:47:41.231409 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mxwtf/must-gather-gbtqv" event={"ID":"e7f569d0-eb50-4458-9144-b8f4c8947bf1","Type":"ContainerStarted","Data":"c4c09f5d02e8d778987bf1b9d67ee56115a1328f9ae89c8890d6ac9e77e66ad6"} Dec 01 09:47:41 crc kubenswrapper[4873]: I1201 09:47:41.270983 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-mxwtf/must-gather-gbtqv" podStartSLOduration=3.27095454 podStartE2EDuration="3.27095454s" podCreationTimestamp="2025-12-01 09:47:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:47:41.259535444 +0000 UTC m=+4037.161643983" watchObservedRunningTime="2025-12-01 09:47:41.27095454 +0000 UTC m=+4037.173063079" Dec 01 09:47:45 crc kubenswrapper[4873]: I1201 09:47:45.209077 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mxwtf/crc-debug-rdg9b"] Dec 01 09:47:45 crc kubenswrapper[4873]: I1201 09:47:45.211391 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mxwtf/crc-debug-rdg9b" Dec 01 09:47:45 crc kubenswrapper[4873]: I1201 09:47:45.290545 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9c9fe293-9f67-441b-a6c6-5624fa898377-host\") pod \"crc-debug-rdg9b\" (UID: \"9c9fe293-9f67-441b-a6c6-5624fa898377\") " pod="openshift-must-gather-mxwtf/crc-debug-rdg9b" Dec 01 09:47:45 crc kubenswrapper[4873]: I1201 09:47:45.290639 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cw5ss\" (UniqueName: \"kubernetes.io/projected/9c9fe293-9f67-441b-a6c6-5624fa898377-kube-api-access-cw5ss\") pod \"crc-debug-rdg9b\" (UID: \"9c9fe293-9f67-441b-a6c6-5624fa898377\") " pod="openshift-must-gather-mxwtf/crc-debug-rdg9b" Dec 01 09:47:45 crc kubenswrapper[4873]: I1201 09:47:45.393357 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9c9fe293-9f67-441b-a6c6-5624fa898377-host\") pod \"crc-debug-rdg9b\" (UID: \"9c9fe293-9f67-441b-a6c6-5624fa898377\") " pod="openshift-must-gather-mxwtf/crc-debug-rdg9b" Dec 01 09:47:45 crc kubenswrapper[4873]: I1201 09:47:45.393838 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cw5ss\" (UniqueName: \"kubernetes.io/projected/9c9fe293-9f67-441b-a6c6-5624fa898377-kube-api-access-cw5ss\") pod \"crc-debug-rdg9b\" (UID: \"9c9fe293-9f67-441b-a6c6-5624fa898377\") " pod="openshift-must-gather-mxwtf/crc-debug-rdg9b" Dec 01 09:47:45 crc kubenswrapper[4873]: I1201 09:47:45.393569 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9c9fe293-9f67-441b-a6c6-5624fa898377-host\") pod \"crc-debug-rdg9b\" (UID: \"9c9fe293-9f67-441b-a6c6-5624fa898377\") " pod="openshift-must-gather-mxwtf/crc-debug-rdg9b" Dec 01 09:47:45 crc kubenswrapper[4873]: I1201 09:47:45.419514 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cw5ss\" (UniqueName: \"kubernetes.io/projected/9c9fe293-9f67-441b-a6c6-5624fa898377-kube-api-access-cw5ss\") pod \"crc-debug-rdg9b\" (UID: \"9c9fe293-9f67-441b-a6c6-5624fa898377\") " pod="openshift-must-gather-mxwtf/crc-debug-rdg9b" Dec 01 09:47:45 crc kubenswrapper[4873]: I1201 09:47:45.537372 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mxwtf/crc-debug-rdg9b" Dec 01 09:47:46 crc kubenswrapper[4873]: I1201 09:47:46.275255 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mxwtf/crc-debug-rdg9b" event={"ID":"9c9fe293-9f67-441b-a6c6-5624fa898377","Type":"ContainerStarted","Data":"0b69ff28cc69f58b8b0993eb581f455022cb677089eea31cc4197b0cc6dbad03"} Dec 01 09:47:46 crc kubenswrapper[4873]: I1201 09:47:46.276008 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mxwtf/crc-debug-rdg9b" event={"ID":"9c9fe293-9f67-441b-a6c6-5624fa898377","Type":"ContainerStarted","Data":"dfb0e9f8c24bf724332150f72678e6f11f592a31255c29322209d89239034542"} Dec 01 09:47:46 crc kubenswrapper[4873]: I1201 09:47:46.297104 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-mxwtf/crc-debug-rdg9b" podStartSLOduration=1.297077679 podStartE2EDuration="1.297077679s" podCreationTimestamp="2025-12-01 09:47:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:47:46.29675515 +0000 UTC m=+4042.198863689" watchObservedRunningTime="2025-12-01 09:47:46.297077679 +0000 UTC m=+4042.199186218" Dec 01 09:48:01 crc kubenswrapper[4873]: I1201 09:48:01.059362 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:48:01 crc kubenswrapper[4873]: I1201 09:48:01.060086 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:48:17 crc kubenswrapper[4873]: I1201 09:48:17.002817 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bfbjm"] Dec 01 09:48:17 crc kubenswrapper[4873]: I1201 09:48:17.007632 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bfbjm" Dec 01 09:48:17 crc kubenswrapper[4873]: I1201 09:48:17.025098 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bfbjm"] Dec 01 09:48:17 crc kubenswrapper[4873]: I1201 09:48:17.119816 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckpgn\" (UniqueName: \"kubernetes.io/projected/18aef89c-bf26-4bd1-a36b-11c311f12a66-kube-api-access-ckpgn\") pod \"community-operators-bfbjm\" (UID: \"18aef89c-bf26-4bd1-a36b-11c311f12a66\") " pod="openshift-marketplace/community-operators-bfbjm" Dec 01 09:48:17 crc kubenswrapper[4873]: I1201 09:48:17.120299 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18aef89c-bf26-4bd1-a36b-11c311f12a66-utilities\") pod \"community-operators-bfbjm\" (UID: \"18aef89c-bf26-4bd1-a36b-11c311f12a66\") " pod="openshift-marketplace/community-operators-bfbjm" Dec 01 09:48:17 crc kubenswrapper[4873]: I1201 09:48:17.120377 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18aef89c-bf26-4bd1-a36b-11c311f12a66-catalog-content\") pod \"community-operators-bfbjm\" (UID: \"18aef89c-bf26-4bd1-a36b-11c311f12a66\") " pod="openshift-marketplace/community-operators-bfbjm" Dec 01 09:48:17 crc kubenswrapper[4873]: I1201 09:48:17.222552 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18aef89c-bf26-4bd1-a36b-11c311f12a66-catalog-content\") pod \"community-operators-bfbjm\" (UID: \"18aef89c-bf26-4bd1-a36b-11c311f12a66\") " pod="openshift-marketplace/community-operators-bfbjm" Dec 01 09:48:17 crc kubenswrapper[4873]: I1201 09:48:17.223044 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckpgn\" (UniqueName: \"kubernetes.io/projected/18aef89c-bf26-4bd1-a36b-11c311f12a66-kube-api-access-ckpgn\") pod \"community-operators-bfbjm\" (UID: \"18aef89c-bf26-4bd1-a36b-11c311f12a66\") " pod="openshift-marketplace/community-operators-bfbjm" Dec 01 09:48:17 crc kubenswrapper[4873]: I1201 09:48:17.223154 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18aef89c-bf26-4bd1-a36b-11c311f12a66-utilities\") pod \"community-operators-bfbjm\" (UID: \"18aef89c-bf26-4bd1-a36b-11c311f12a66\") " pod="openshift-marketplace/community-operators-bfbjm" Dec 01 09:48:17 crc kubenswrapper[4873]: I1201 09:48:17.223296 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18aef89c-bf26-4bd1-a36b-11c311f12a66-catalog-content\") pod \"community-operators-bfbjm\" (UID: \"18aef89c-bf26-4bd1-a36b-11c311f12a66\") " pod="openshift-marketplace/community-operators-bfbjm" Dec 01 09:48:17 crc kubenswrapper[4873]: I1201 09:48:17.223685 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18aef89c-bf26-4bd1-a36b-11c311f12a66-utilities\") pod \"community-operators-bfbjm\" (UID: \"18aef89c-bf26-4bd1-a36b-11c311f12a66\") " pod="openshift-marketplace/community-operators-bfbjm" Dec 01 09:48:17 crc kubenswrapper[4873]: I1201 09:48:17.252114 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckpgn\" (UniqueName: \"kubernetes.io/projected/18aef89c-bf26-4bd1-a36b-11c311f12a66-kube-api-access-ckpgn\") pod \"community-operators-bfbjm\" (UID: \"18aef89c-bf26-4bd1-a36b-11c311f12a66\") " pod="openshift-marketplace/community-operators-bfbjm" Dec 01 09:48:17 crc kubenswrapper[4873]: I1201 09:48:17.332329 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bfbjm" Dec 01 09:48:17 crc kubenswrapper[4873]: I1201 09:48:17.915443 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bfbjm"] Dec 01 09:48:18 crc kubenswrapper[4873]: I1201 09:48:18.621296 4873 generic.go:334] "Generic (PLEG): container finished" podID="18aef89c-bf26-4bd1-a36b-11c311f12a66" containerID="b595bbe8475b825da954e09b451fbce58cf605b91264690759ba06796fcced4e" exitCode=0 Dec 01 09:48:18 crc kubenswrapper[4873]: I1201 09:48:18.621354 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bfbjm" event={"ID":"18aef89c-bf26-4bd1-a36b-11c311f12a66","Type":"ContainerDied","Data":"b595bbe8475b825da954e09b451fbce58cf605b91264690759ba06796fcced4e"} Dec 01 09:48:18 crc kubenswrapper[4873]: I1201 09:48:18.623005 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bfbjm" event={"ID":"18aef89c-bf26-4bd1-a36b-11c311f12a66","Type":"ContainerStarted","Data":"0479ab4e8432990640dd65020a6923bc1d8631febf661b6d846a9784f44611f5"} Dec 01 09:48:18 crc kubenswrapper[4873]: I1201 09:48:18.623816 4873 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 09:48:19 crc kubenswrapper[4873]: I1201 09:48:19.634240 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bfbjm" event={"ID":"18aef89c-bf26-4bd1-a36b-11c311f12a66","Type":"ContainerStarted","Data":"646b636a94c75853c641aef581a2c65dbad673890f404955f67194edc08b760c"} Dec 01 09:48:20 crc kubenswrapper[4873]: E1201 09:48:20.292091 4873 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod18aef89c_bf26_4bd1_a36b_11c311f12a66.slice/crio-646b636a94c75853c641aef581a2c65dbad673890f404955f67194edc08b760c.scope\": RecentStats: unable to find data in memory cache]" Dec 01 09:48:20 crc kubenswrapper[4873]: I1201 09:48:20.655040 4873 generic.go:334] "Generic (PLEG): container finished" podID="18aef89c-bf26-4bd1-a36b-11c311f12a66" containerID="646b636a94c75853c641aef581a2c65dbad673890f404955f67194edc08b760c" exitCode=0 Dec 01 09:48:20 crc kubenswrapper[4873]: I1201 09:48:20.655104 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bfbjm" event={"ID":"18aef89c-bf26-4bd1-a36b-11c311f12a66","Type":"ContainerDied","Data":"646b636a94c75853c641aef581a2c65dbad673890f404955f67194edc08b760c"} Dec 01 09:48:21 crc kubenswrapper[4873]: I1201 09:48:21.668535 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bfbjm" event={"ID":"18aef89c-bf26-4bd1-a36b-11c311f12a66","Type":"ContainerStarted","Data":"3eccd6b42ddd633e2507e6779efa17f88855ffe9504d4aa37373edbc3987a9ac"} Dec 01 09:48:21 crc kubenswrapper[4873]: I1201 09:48:21.691392 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bfbjm" podStartSLOduration=2.896724747 podStartE2EDuration="5.691370297s" podCreationTimestamp="2025-12-01 09:48:16 +0000 UTC" firstStartedPulling="2025-12-01 09:48:18.623494323 +0000 UTC m=+4074.525602872" lastFinishedPulling="2025-12-01 09:48:21.418139883 +0000 UTC m=+4077.320248422" observedRunningTime="2025-12-01 09:48:21.689274813 +0000 UTC m=+4077.591383352" watchObservedRunningTime="2025-12-01 09:48:21.691370297 +0000 UTC m=+4077.593478836" Dec 01 09:48:23 crc kubenswrapper[4873]: I1201 09:48:23.690832 4873 generic.go:334] "Generic (PLEG): container finished" podID="9c9fe293-9f67-441b-a6c6-5624fa898377" containerID="0b69ff28cc69f58b8b0993eb581f455022cb677089eea31cc4197b0cc6dbad03" exitCode=0 Dec 01 09:48:23 crc kubenswrapper[4873]: I1201 09:48:23.691101 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mxwtf/crc-debug-rdg9b" event={"ID":"9c9fe293-9f67-441b-a6c6-5624fa898377","Type":"ContainerDied","Data":"0b69ff28cc69f58b8b0993eb581f455022cb677089eea31cc4197b0cc6dbad03"} Dec 01 09:48:24 crc kubenswrapper[4873]: I1201 09:48:24.815485 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mxwtf/crc-debug-rdg9b" Dec 01 09:48:24 crc kubenswrapper[4873]: I1201 09:48:24.860830 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mxwtf/crc-debug-rdg9b"] Dec 01 09:48:24 crc kubenswrapper[4873]: I1201 09:48:24.869921 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mxwtf/crc-debug-rdg9b"] Dec 01 09:48:24 crc kubenswrapper[4873]: I1201 09:48:24.932461 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cw5ss\" (UniqueName: \"kubernetes.io/projected/9c9fe293-9f67-441b-a6c6-5624fa898377-kube-api-access-cw5ss\") pod \"9c9fe293-9f67-441b-a6c6-5624fa898377\" (UID: \"9c9fe293-9f67-441b-a6c6-5624fa898377\") " Dec 01 09:48:24 crc kubenswrapper[4873]: I1201 09:48:24.932775 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9c9fe293-9f67-441b-a6c6-5624fa898377-host\") pod \"9c9fe293-9f67-441b-a6c6-5624fa898377\" (UID: \"9c9fe293-9f67-441b-a6c6-5624fa898377\") " Dec 01 09:48:24 crc kubenswrapper[4873]: I1201 09:48:24.932933 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9c9fe293-9f67-441b-a6c6-5624fa898377-host" (OuterVolumeSpecName: "host") pod "9c9fe293-9f67-441b-a6c6-5624fa898377" (UID: "9c9fe293-9f67-441b-a6c6-5624fa898377"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:48:24 crc kubenswrapper[4873]: I1201 09:48:24.933601 4873 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9c9fe293-9f67-441b-a6c6-5624fa898377-host\") on node \"crc\" DevicePath \"\"" Dec 01 09:48:24 crc kubenswrapper[4873]: I1201 09:48:24.945524 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c9fe293-9f67-441b-a6c6-5624fa898377-kube-api-access-cw5ss" (OuterVolumeSpecName: "kube-api-access-cw5ss") pod "9c9fe293-9f67-441b-a6c6-5624fa898377" (UID: "9c9fe293-9f67-441b-a6c6-5624fa898377"). InnerVolumeSpecName "kube-api-access-cw5ss". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:48:25 crc kubenswrapper[4873]: I1201 09:48:25.035707 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cw5ss\" (UniqueName: \"kubernetes.io/projected/9c9fe293-9f67-441b-a6c6-5624fa898377-kube-api-access-cw5ss\") on node \"crc\" DevicePath \"\"" Dec 01 09:48:25 crc kubenswrapper[4873]: I1201 09:48:25.713245 4873 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dfb0e9f8c24bf724332150f72678e6f11f592a31255c29322209d89239034542" Dec 01 09:48:25 crc kubenswrapper[4873]: I1201 09:48:25.713298 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mxwtf/crc-debug-rdg9b" Dec 01 09:48:26 crc kubenswrapper[4873]: I1201 09:48:26.042948 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mxwtf/crc-debug-d26gc"] Dec 01 09:48:26 crc kubenswrapper[4873]: E1201 09:48:26.045641 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c9fe293-9f67-441b-a6c6-5624fa898377" containerName="container-00" Dec 01 09:48:26 crc kubenswrapper[4873]: I1201 09:48:26.045668 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c9fe293-9f67-441b-a6c6-5624fa898377" containerName="container-00" Dec 01 09:48:26 crc kubenswrapper[4873]: I1201 09:48:26.045882 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c9fe293-9f67-441b-a6c6-5624fa898377" containerName="container-00" Dec 01 09:48:26 crc kubenswrapper[4873]: I1201 09:48:26.046865 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mxwtf/crc-debug-d26gc" Dec 01 09:48:26 crc kubenswrapper[4873]: I1201 09:48:26.058773 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a672e380-526f-42d6-ada5-83073fa55537-host\") pod \"crc-debug-d26gc\" (UID: \"a672e380-526f-42d6-ada5-83073fa55537\") " pod="openshift-must-gather-mxwtf/crc-debug-d26gc" Dec 01 09:48:26 crc kubenswrapper[4873]: I1201 09:48:26.059223 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28zmd\" (UniqueName: \"kubernetes.io/projected/a672e380-526f-42d6-ada5-83073fa55537-kube-api-access-28zmd\") pod \"crc-debug-d26gc\" (UID: \"a672e380-526f-42d6-ada5-83073fa55537\") " pod="openshift-must-gather-mxwtf/crc-debug-d26gc" Dec 01 09:48:26 crc kubenswrapper[4873]: I1201 09:48:26.162142 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a672e380-526f-42d6-ada5-83073fa55537-host\") pod \"crc-debug-d26gc\" (UID: \"a672e380-526f-42d6-ada5-83073fa55537\") " pod="openshift-must-gather-mxwtf/crc-debug-d26gc" Dec 01 09:48:26 crc kubenswrapper[4873]: I1201 09:48:26.162270 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28zmd\" (UniqueName: \"kubernetes.io/projected/a672e380-526f-42d6-ada5-83073fa55537-kube-api-access-28zmd\") pod \"crc-debug-d26gc\" (UID: \"a672e380-526f-42d6-ada5-83073fa55537\") " pod="openshift-must-gather-mxwtf/crc-debug-d26gc" Dec 01 09:48:26 crc kubenswrapper[4873]: I1201 09:48:26.162300 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a672e380-526f-42d6-ada5-83073fa55537-host\") pod \"crc-debug-d26gc\" (UID: \"a672e380-526f-42d6-ada5-83073fa55537\") " pod="openshift-must-gather-mxwtf/crc-debug-d26gc" Dec 01 09:48:26 crc kubenswrapper[4873]: I1201 09:48:26.185978 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28zmd\" (UniqueName: \"kubernetes.io/projected/a672e380-526f-42d6-ada5-83073fa55537-kube-api-access-28zmd\") pod \"crc-debug-d26gc\" (UID: \"a672e380-526f-42d6-ada5-83073fa55537\") " pod="openshift-must-gather-mxwtf/crc-debug-d26gc" Dec 01 09:48:26 crc kubenswrapper[4873]: I1201 09:48:26.368608 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mxwtf/crc-debug-d26gc" Dec 01 09:48:26 crc kubenswrapper[4873]: I1201 09:48:26.443448 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c9fe293-9f67-441b-a6c6-5624fa898377" path="/var/lib/kubelet/pods/9c9fe293-9f67-441b-a6c6-5624fa898377/volumes" Dec 01 09:48:26 crc kubenswrapper[4873]: I1201 09:48:26.725842 4873 generic.go:334] "Generic (PLEG): container finished" podID="a672e380-526f-42d6-ada5-83073fa55537" containerID="597cb8b3cc8cacf93168e49d8dd99a8b7c925c53c9ad4cc4c3c20c4eb050e55f" exitCode=0 Dec 01 09:48:26 crc kubenswrapper[4873]: I1201 09:48:26.725924 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mxwtf/crc-debug-d26gc" event={"ID":"a672e380-526f-42d6-ada5-83073fa55537","Type":"ContainerDied","Data":"597cb8b3cc8cacf93168e49d8dd99a8b7c925c53c9ad4cc4c3c20c4eb050e55f"} Dec 01 09:48:26 crc kubenswrapper[4873]: I1201 09:48:26.726189 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mxwtf/crc-debug-d26gc" event={"ID":"a672e380-526f-42d6-ada5-83073fa55537","Type":"ContainerStarted","Data":"c7eeca3714eca3dcbc558f918627beb019d13f8a751b2f29d22d5e80ef3ab850"} Dec 01 09:48:27 crc kubenswrapper[4873]: I1201 09:48:27.256702 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mxwtf/crc-debug-d26gc"] Dec 01 09:48:27 crc kubenswrapper[4873]: I1201 09:48:27.265320 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mxwtf/crc-debug-d26gc"] Dec 01 09:48:27 crc kubenswrapper[4873]: I1201 09:48:27.332492 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bfbjm" Dec 01 09:48:27 crc kubenswrapper[4873]: I1201 09:48:27.332542 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bfbjm" Dec 01 09:48:27 crc kubenswrapper[4873]: I1201 09:48:27.397448 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bfbjm" Dec 01 09:48:27 crc kubenswrapper[4873]: I1201 09:48:27.804644 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bfbjm" Dec 01 09:48:27 crc kubenswrapper[4873]: I1201 09:48:27.866945 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bfbjm"] Dec 01 09:48:27 crc kubenswrapper[4873]: I1201 09:48:27.877903 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mxwtf/crc-debug-d26gc" Dec 01 09:48:27 crc kubenswrapper[4873]: I1201 09:48:27.904831 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28zmd\" (UniqueName: \"kubernetes.io/projected/a672e380-526f-42d6-ada5-83073fa55537-kube-api-access-28zmd\") pod \"a672e380-526f-42d6-ada5-83073fa55537\" (UID: \"a672e380-526f-42d6-ada5-83073fa55537\") " Dec 01 09:48:27 crc kubenswrapper[4873]: I1201 09:48:27.904949 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a672e380-526f-42d6-ada5-83073fa55537-host\") pod \"a672e380-526f-42d6-ada5-83073fa55537\" (UID: \"a672e380-526f-42d6-ada5-83073fa55537\") " Dec 01 09:48:27 crc kubenswrapper[4873]: I1201 09:48:27.905622 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a672e380-526f-42d6-ada5-83073fa55537-host" (OuterVolumeSpecName: "host") pod "a672e380-526f-42d6-ada5-83073fa55537" (UID: "a672e380-526f-42d6-ada5-83073fa55537"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:48:27 crc kubenswrapper[4873]: I1201 09:48:27.914505 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a672e380-526f-42d6-ada5-83073fa55537-kube-api-access-28zmd" (OuterVolumeSpecName: "kube-api-access-28zmd") pod "a672e380-526f-42d6-ada5-83073fa55537" (UID: "a672e380-526f-42d6-ada5-83073fa55537"). InnerVolumeSpecName "kube-api-access-28zmd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:48:28 crc kubenswrapper[4873]: I1201 09:48:28.006464 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28zmd\" (UniqueName: \"kubernetes.io/projected/a672e380-526f-42d6-ada5-83073fa55537-kube-api-access-28zmd\") on node \"crc\" DevicePath \"\"" Dec 01 09:48:28 crc kubenswrapper[4873]: I1201 09:48:28.006503 4873 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a672e380-526f-42d6-ada5-83073fa55537-host\") on node \"crc\" DevicePath \"\"" Dec 01 09:48:28 crc kubenswrapper[4873]: I1201 09:48:28.457817 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a672e380-526f-42d6-ada5-83073fa55537" path="/var/lib/kubelet/pods/a672e380-526f-42d6-ada5-83073fa55537/volumes" Dec 01 09:48:28 crc kubenswrapper[4873]: I1201 09:48:28.492619 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mxwtf/crc-debug-dtpwt"] Dec 01 09:48:28 crc kubenswrapper[4873]: E1201 09:48:28.493754 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a672e380-526f-42d6-ada5-83073fa55537" containerName="container-00" Dec 01 09:48:28 crc kubenswrapper[4873]: I1201 09:48:28.494073 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="a672e380-526f-42d6-ada5-83073fa55537" containerName="container-00" Dec 01 09:48:28 crc kubenswrapper[4873]: I1201 09:48:28.494579 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="a672e380-526f-42d6-ada5-83073fa55537" containerName="container-00" Dec 01 09:48:28 crc kubenswrapper[4873]: I1201 09:48:28.496318 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mxwtf/crc-debug-dtpwt" Dec 01 09:48:28 crc kubenswrapper[4873]: I1201 09:48:28.527194 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtfjl\" (UniqueName: \"kubernetes.io/projected/bea34349-b8c1-4e5b-b2bb-839fcd0eb76d-kube-api-access-rtfjl\") pod \"crc-debug-dtpwt\" (UID: \"bea34349-b8c1-4e5b-b2bb-839fcd0eb76d\") " pod="openshift-must-gather-mxwtf/crc-debug-dtpwt" Dec 01 09:48:28 crc kubenswrapper[4873]: I1201 09:48:28.527632 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bea34349-b8c1-4e5b-b2bb-839fcd0eb76d-host\") pod \"crc-debug-dtpwt\" (UID: \"bea34349-b8c1-4e5b-b2bb-839fcd0eb76d\") " pod="openshift-must-gather-mxwtf/crc-debug-dtpwt" Dec 01 09:48:28 crc kubenswrapper[4873]: I1201 09:48:28.630232 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bea34349-b8c1-4e5b-b2bb-839fcd0eb76d-host\") pod \"crc-debug-dtpwt\" (UID: \"bea34349-b8c1-4e5b-b2bb-839fcd0eb76d\") " pod="openshift-must-gather-mxwtf/crc-debug-dtpwt" Dec 01 09:48:28 crc kubenswrapper[4873]: I1201 09:48:28.630417 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtfjl\" (UniqueName: \"kubernetes.io/projected/bea34349-b8c1-4e5b-b2bb-839fcd0eb76d-kube-api-access-rtfjl\") pod \"crc-debug-dtpwt\" (UID: \"bea34349-b8c1-4e5b-b2bb-839fcd0eb76d\") " pod="openshift-must-gather-mxwtf/crc-debug-dtpwt" Dec 01 09:48:28 crc kubenswrapper[4873]: I1201 09:48:28.631308 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bea34349-b8c1-4e5b-b2bb-839fcd0eb76d-host\") pod \"crc-debug-dtpwt\" (UID: \"bea34349-b8c1-4e5b-b2bb-839fcd0eb76d\") " pod="openshift-must-gather-mxwtf/crc-debug-dtpwt" Dec 01 09:48:28 crc kubenswrapper[4873]: I1201 09:48:28.666833 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtfjl\" (UniqueName: \"kubernetes.io/projected/bea34349-b8c1-4e5b-b2bb-839fcd0eb76d-kube-api-access-rtfjl\") pod \"crc-debug-dtpwt\" (UID: \"bea34349-b8c1-4e5b-b2bb-839fcd0eb76d\") " pod="openshift-must-gather-mxwtf/crc-debug-dtpwt" Dec 01 09:48:28 crc kubenswrapper[4873]: I1201 09:48:28.750028 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mxwtf/crc-debug-d26gc" Dec 01 09:48:28 crc kubenswrapper[4873]: I1201 09:48:28.750052 4873 scope.go:117] "RemoveContainer" containerID="597cb8b3cc8cacf93168e49d8dd99a8b7c925c53c9ad4cc4c3c20c4eb050e55f" Dec 01 09:48:28 crc kubenswrapper[4873]: I1201 09:48:28.820838 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mxwtf/crc-debug-dtpwt" Dec 01 09:48:28 crc kubenswrapper[4873]: W1201 09:48:28.852351 4873 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbea34349_b8c1_4e5b_b2bb_839fcd0eb76d.slice/crio-5caad50f5200f3bfe87547aca28ad138a690b635f31b33896c7f7c1eff6c8f85 WatchSource:0}: Error finding container 5caad50f5200f3bfe87547aca28ad138a690b635f31b33896c7f7c1eff6c8f85: Status 404 returned error can't find the container with id 5caad50f5200f3bfe87547aca28ad138a690b635f31b33896c7f7c1eff6c8f85 Dec 01 09:48:29 crc kubenswrapper[4873]: I1201 09:48:29.762356 4873 generic.go:334] "Generic (PLEG): container finished" podID="bea34349-b8c1-4e5b-b2bb-839fcd0eb76d" containerID="3069cc8f67795457e00c6dc72c387145628bd9c5b96ad410d72fe4288b9ee030" exitCode=0 Dec 01 09:48:29 crc kubenswrapper[4873]: I1201 09:48:29.762456 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mxwtf/crc-debug-dtpwt" event={"ID":"bea34349-b8c1-4e5b-b2bb-839fcd0eb76d","Type":"ContainerDied","Data":"3069cc8f67795457e00c6dc72c387145628bd9c5b96ad410d72fe4288b9ee030"} Dec 01 09:48:29 crc kubenswrapper[4873]: I1201 09:48:29.762779 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mxwtf/crc-debug-dtpwt" event={"ID":"bea34349-b8c1-4e5b-b2bb-839fcd0eb76d","Type":"ContainerStarted","Data":"5caad50f5200f3bfe87547aca28ad138a690b635f31b33896c7f7c1eff6c8f85"} Dec 01 09:48:29 crc kubenswrapper[4873]: I1201 09:48:29.765313 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-bfbjm" podUID="18aef89c-bf26-4bd1-a36b-11c311f12a66" containerName="registry-server" containerID="cri-o://3eccd6b42ddd633e2507e6779efa17f88855ffe9504d4aa37373edbc3987a9ac" gracePeriod=2 Dec 01 09:48:29 crc kubenswrapper[4873]: I1201 09:48:29.814357 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mxwtf/crc-debug-dtpwt"] Dec 01 09:48:29 crc kubenswrapper[4873]: I1201 09:48:29.827967 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mxwtf/crc-debug-dtpwt"] Dec 01 09:48:30 crc kubenswrapper[4873]: I1201 09:48:30.297003 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bfbjm" Dec 01 09:48:30 crc kubenswrapper[4873]: I1201 09:48:30.379755 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ckpgn\" (UniqueName: \"kubernetes.io/projected/18aef89c-bf26-4bd1-a36b-11c311f12a66-kube-api-access-ckpgn\") pod \"18aef89c-bf26-4bd1-a36b-11c311f12a66\" (UID: \"18aef89c-bf26-4bd1-a36b-11c311f12a66\") " Dec 01 09:48:30 crc kubenswrapper[4873]: I1201 09:48:30.379850 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18aef89c-bf26-4bd1-a36b-11c311f12a66-catalog-content\") pod \"18aef89c-bf26-4bd1-a36b-11c311f12a66\" (UID: \"18aef89c-bf26-4bd1-a36b-11c311f12a66\") " Dec 01 09:48:30 crc kubenswrapper[4873]: I1201 09:48:30.380007 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18aef89c-bf26-4bd1-a36b-11c311f12a66-utilities\") pod \"18aef89c-bf26-4bd1-a36b-11c311f12a66\" (UID: \"18aef89c-bf26-4bd1-a36b-11c311f12a66\") " Dec 01 09:48:30 crc kubenswrapper[4873]: I1201 09:48:30.381325 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18aef89c-bf26-4bd1-a36b-11c311f12a66-utilities" (OuterVolumeSpecName: "utilities") pod "18aef89c-bf26-4bd1-a36b-11c311f12a66" (UID: "18aef89c-bf26-4bd1-a36b-11c311f12a66"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:48:30 crc kubenswrapper[4873]: I1201 09:48:30.388358 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18aef89c-bf26-4bd1-a36b-11c311f12a66-kube-api-access-ckpgn" (OuterVolumeSpecName: "kube-api-access-ckpgn") pod "18aef89c-bf26-4bd1-a36b-11c311f12a66" (UID: "18aef89c-bf26-4bd1-a36b-11c311f12a66"). InnerVolumeSpecName "kube-api-access-ckpgn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:48:30 crc kubenswrapper[4873]: I1201 09:48:30.448143 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18aef89c-bf26-4bd1-a36b-11c311f12a66-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "18aef89c-bf26-4bd1-a36b-11c311f12a66" (UID: "18aef89c-bf26-4bd1-a36b-11c311f12a66"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:48:30 crc kubenswrapper[4873]: I1201 09:48:30.482799 4873 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18aef89c-bf26-4bd1-a36b-11c311f12a66-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 09:48:30 crc kubenswrapper[4873]: I1201 09:48:30.482829 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ckpgn\" (UniqueName: \"kubernetes.io/projected/18aef89c-bf26-4bd1-a36b-11c311f12a66-kube-api-access-ckpgn\") on node \"crc\" DevicePath \"\"" Dec 01 09:48:30 crc kubenswrapper[4873]: I1201 09:48:30.482840 4873 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18aef89c-bf26-4bd1-a36b-11c311f12a66-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 09:48:30 crc kubenswrapper[4873]: I1201 09:48:30.777231 4873 generic.go:334] "Generic (PLEG): container finished" podID="18aef89c-bf26-4bd1-a36b-11c311f12a66" containerID="3eccd6b42ddd633e2507e6779efa17f88855ffe9504d4aa37373edbc3987a9ac" exitCode=0 Dec 01 09:48:30 crc kubenswrapper[4873]: I1201 09:48:30.777312 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bfbjm" Dec 01 09:48:30 crc kubenswrapper[4873]: I1201 09:48:30.777335 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bfbjm" event={"ID":"18aef89c-bf26-4bd1-a36b-11c311f12a66","Type":"ContainerDied","Data":"3eccd6b42ddd633e2507e6779efa17f88855ffe9504d4aa37373edbc3987a9ac"} Dec 01 09:48:30 crc kubenswrapper[4873]: I1201 09:48:30.777840 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bfbjm" event={"ID":"18aef89c-bf26-4bd1-a36b-11c311f12a66","Type":"ContainerDied","Data":"0479ab4e8432990640dd65020a6923bc1d8631febf661b6d846a9784f44611f5"} Dec 01 09:48:30 crc kubenswrapper[4873]: I1201 09:48:30.777874 4873 scope.go:117] "RemoveContainer" containerID="3eccd6b42ddd633e2507e6779efa17f88855ffe9504d4aa37373edbc3987a9ac" Dec 01 09:48:30 crc kubenswrapper[4873]: I1201 09:48:30.854997 4873 scope.go:117] "RemoveContainer" containerID="646b636a94c75853c641aef581a2c65dbad673890f404955f67194edc08b760c" Dec 01 09:48:30 crc kubenswrapper[4873]: I1201 09:48:30.858389 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mxwtf/crc-debug-dtpwt" Dec 01 09:48:30 crc kubenswrapper[4873]: I1201 09:48:30.890442 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rtfjl\" (UniqueName: \"kubernetes.io/projected/bea34349-b8c1-4e5b-b2bb-839fcd0eb76d-kube-api-access-rtfjl\") pod \"bea34349-b8c1-4e5b-b2bb-839fcd0eb76d\" (UID: \"bea34349-b8c1-4e5b-b2bb-839fcd0eb76d\") " Dec 01 09:48:30 crc kubenswrapper[4873]: I1201 09:48:30.890519 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bea34349-b8c1-4e5b-b2bb-839fcd0eb76d-host\") pod \"bea34349-b8c1-4e5b-b2bb-839fcd0eb76d\" (UID: \"bea34349-b8c1-4e5b-b2bb-839fcd0eb76d\") " Dec 01 09:48:30 crc kubenswrapper[4873]: I1201 09:48:30.890654 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bea34349-b8c1-4e5b-b2bb-839fcd0eb76d-host" (OuterVolumeSpecName: "host") pod "bea34349-b8c1-4e5b-b2bb-839fcd0eb76d" (UID: "bea34349-b8c1-4e5b-b2bb-839fcd0eb76d"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:48:30 crc kubenswrapper[4873]: I1201 09:48:30.891659 4873 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bea34349-b8c1-4e5b-b2bb-839fcd0eb76d-host\") on node \"crc\" DevicePath \"\"" Dec 01 09:48:30 crc kubenswrapper[4873]: I1201 09:48:30.897407 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bfbjm"] Dec 01 09:48:30 crc kubenswrapper[4873]: I1201 09:48:30.897643 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bea34349-b8c1-4e5b-b2bb-839fcd0eb76d-kube-api-access-rtfjl" (OuterVolumeSpecName: "kube-api-access-rtfjl") pod "bea34349-b8c1-4e5b-b2bb-839fcd0eb76d" (UID: "bea34349-b8c1-4e5b-b2bb-839fcd0eb76d"). InnerVolumeSpecName "kube-api-access-rtfjl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:48:30 crc kubenswrapper[4873]: I1201 09:48:30.898317 4873 scope.go:117] "RemoveContainer" containerID="b595bbe8475b825da954e09b451fbce58cf605b91264690759ba06796fcced4e" Dec 01 09:48:30 crc kubenswrapper[4873]: I1201 09:48:30.913676 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-bfbjm"] Dec 01 09:48:30 crc kubenswrapper[4873]: I1201 09:48:30.993561 4873 scope.go:117] "RemoveContainer" containerID="3eccd6b42ddd633e2507e6779efa17f88855ffe9504d4aa37373edbc3987a9ac" Dec 01 09:48:30 crc kubenswrapper[4873]: E1201 09:48:30.994090 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3eccd6b42ddd633e2507e6779efa17f88855ffe9504d4aa37373edbc3987a9ac\": container with ID starting with 3eccd6b42ddd633e2507e6779efa17f88855ffe9504d4aa37373edbc3987a9ac not found: ID does not exist" containerID="3eccd6b42ddd633e2507e6779efa17f88855ffe9504d4aa37373edbc3987a9ac" Dec 01 09:48:30 crc kubenswrapper[4873]: I1201 09:48:30.994163 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3eccd6b42ddd633e2507e6779efa17f88855ffe9504d4aa37373edbc3987a9ac"} err="failed to get container status \"3eccd6b42ddd633e2507e6779efa17f88855ffe9504d4aa37373edbc3987a9ac\": rpc error: code = NotFound desc = could not find container \"3eccd6b42ddd633e2507e6779efa17f88855ffe9504d4aa37373edbc3987a9ac\": container with ID starting with 3eccd6b42ddd633e2507e6779efa17f88855ffe9504d4aa37373edbc3987a9ac not found: ID does not exist" Dec 01 09:48:30 crc kubenswrapper[4873]: I1201 09:48:30.994200 4873 scope.go:117] "RemoveContainer" containerID="646b636a94c75853c641aef581a2c65dbad673890f404955f67194edc08b760c" Dec 01 09:48:30 crc kubenswrapper[4873]: E1201 09:48:30.994586 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"646b636a94c75853c641aef581a2c65dbad673890f404955f67194edc08b760c\": container with ID starting with 646b636a94c75853c641aef581a2c65dbad673890f404955f67194edc08b760c not found: ID does not exist" containerID="646b636a94c75853c641aef581a2c65dbad673890f404955f67194edc08b760c" Dec 01 09:48:30 crc kubenswrapper[4873]: I1201 09:48:30.994639 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"646b636a94c75853c641aef581a2c65dbad673890f404955f67194edc08b760c"} err="failed to get container status \"646b636a94c75853c641aef581a2c65dbad673890f404955f67194edc08b760c\": rpc error: code = NotFound desc = could not find container \"646b636a94c75853c641aef581a2c65dbad673890f404955f67194edc08b760c\": container with ID starting with 646b636a94c75853c641aef581a2c65dbad673890f404955f67194edc08b760c not found: ID does not exist" Dec 01 09:48:30 crc kubenswrapper[4873]: I1201 09:48:30.994675 4873 scope.go:117] "RemoveContainer" containerID="b595bbe8475b825da954e09b451fbce58cf605b91264690759ba06796fcced4e" Dec 01 09:48:30 crc kubenswrapper[4873]: E1201 09:48:30.995097 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b595bbe8475b825da954e09b451fbce58cf605b91264690759ba06796fcced4e\": container with ID starting with b595bbe8475b825da954e09b451fbce58cf605b91264690759ba06796fcced4e not found: ID does not exist" containerID="b595bbe8475b825da954e09b451fbce58cf605b91264690759ba06796fcced4e" Dec 01 09:48:30 crc kubenswrapper[4873]: I1201 09:48:30.995178 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b595bbe8475b825da954e09b451fbce58cf605b91264690759ba06796fcced4e"} err="failed to get container status \"b595bbe8475b825da954e09b451fbce58cf605b91264690759ba06796fcced4e\": rpc error: code = NotFound desc = could not find container \"b595bbe8475b825da954e09b451fbce58cf605b91264690759ba06796fcced4e\": container with ID starting with b595bbe8475b825da954e09b451fbce58cf605b91264690759ba06796fcced4e not found: ID does not exist" Dec 01 09:48:30 crc kubenswrapper[4873]: I1201 09:48:30.995374 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rtfjl\" (UniqueName: \"kubernetes.io/projected/bea34349-b8c1-4e5b-b2bb-839fcd0eb76d-kube-api-access-rtfjl\") on node \"crc\" DevicePath \"\"" Dec 01 09:48:31 crc kubenswrapper[4873]: I1201 09:48:31.059772 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:48:31 crc kubenswrapper[4873]: I1201 09:48:31.060496 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:48:31 crc kubenswrapper[4873]: I1201 09:48:31.060574 4873 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" Dec 01 09:48:31 crc kubenswrapper[4873]: I1201 09:48:31.061894 4873 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"29490d89d4ce9ae938873b01532387c61bd2469f197fb97051e2685e7e6f8a4e"} pod="openshift-machine-config-operator/machine-config-daemon-scwpp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 09:48:31 crc kubenswrapper[4873]: I1201 09:48:31.062005 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" containerID="cri-o://29490d89d4ce9ae938873b01532387c61bd2469f197fb97051e2685e7e6f8a4e" gracePeriod=600 Dec 01 09:48:31 crc kubenswrapper[4873]: I1201 09:48:31.789397 4873 generic.go:334] "Generic (PLEG): container finished" podID="fef7b114-0e07-402d-a37b-315c36011f4b" containerID="29490d89d4ce9ae938873b01532387c61bd2469f197fb97051e2685e7e6f8a4e" exitCode=0 Dec 01 09:48:31 crc kubenswrapper[4873]: I1201 09:48:31.789522 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" event={"ID":"fef7b114-0e07-402d-a37b-315c36011f4b","Type":"ContainerDied","Data":"29490d89d4ce9ae938873b01532387c61bd2469f197fb97051e2685e7e6f8a4e"} Dec 01 09:48:31 crc kubenswrapper[4873]: I1201 09:48:31.789576 4873 scope.go:117] "RemoveContainer" containerID="c101914cc0d7462afe57653cc6d66f5335bbd707b5b5acec3231122718624d93" Dec 01 09:48:31 crc kubenswrapper[4873]: I1201 09:48:31.793533 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mxwtf/crc-debug-dtpwt" Dec 01 09:48:31 crc kubenswrapper[4873]: I1201 09:48:31.830395 4873 scope.go:117] "RemoveContainer" containerID="3069cc8f67795457e00c6dc72c387145628bd9c5b96ad410d72fe4288b9ee030" Dec 01 09:48:32 crc kubenswrapper[4873]: I1201 09:48:32.441310 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18aef89c-bf26-4bd1-a36b-11c311f12a66" path="/var/lib/kubelet/pods/18aef89c-bf26-4bd1-a36b-11c311f12a66/volumes" Dec 01 09:48:32 crc kubenswrapper[4873]: I1201 09:48:32.445557 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bea34349-b8c1-4e5b-b2bb-839fcd0eb76d" path="/var/lib/kubelet/pods/bea34349-b8c1-4e5b-b2bb-839fcd0eb76d/volumes" Dec 01 09:48:32 crc kubenswrapper[4873]: I1201 09:48:32.807176 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" event={"ID":"fef7b114-0e07-402d-a37b-315c36011f4b","Type":"ContainerStarted","Data":"67abbbfa1eead10a18ea3d3104e50a98c59b5983a7e8d4be64c2d03f882521a7"} Dec 01 09:48:50 crc kubenswrapper[4873]: I1201 09:48:50.758358 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-s826v"] Dec 01 09:48:50 crc kubenswrapper[4873]: E1201 09:48:50.759417 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18aef89c-bf26-4bd1-a36b-11c311f12a66" containerName="extract-content" Dec 01 09:48:50 crc kubenswrapper[4873]: I1201 09:48:50.759433 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="18aef89c-bf26-4bd1-a36b-11c311f12a66" containerName="extract-content" Dec 01 09:48:50 crc kubenswrapper[4873]: E1201 09:48:50.759451 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bea34349-b8c1-4e5b-b2bb-839fcd0eb76d" containerName="container-00" Dec 01 09:48:50 crc kubenswrapper[4873]: I1201 09:48:50.759457 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="bea34349-b8c1-4e5b-b2bb-839fcd0eb76d" containerName="container-00" Dec 01 09:48:50 crc kubenswrapper[4873]: E1201 09:48:50.759484 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18aef89c-bf26-4bd1-a36b-11c311f12a66" containerName="extract-utilities" Dec 01 09:48:50 crc kubenswrapper[4873]: I1201 09:48:50.759491 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="18aef89c-bf26-4bd1-a36b-11c311f12a66" containerName="extract-utilities" Dec 01 09:48:50 crc kubenswrapper[4873]: E1201 09:48:50.759506 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18aef89c-bf26-4bd1-a36b-11c311f12a66" containerName="registry-server" Dec 01 09:48:50 crc kubenswrapper[4873]: I1201 09:48:50.759511 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="18aef89c-bf26-4bd1-a36b-11c311f12a66" containerName="registry-server" Dec 01 09:48:50 crc kubenswrapper[4873]: I1201 09:48:50.759753 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="18aef89c-bf26-4bd1-a36b-11c311f12a66" containerName="registry-server" Dec 01 09:48:50 crc kubenswrapper[4873]: I1201 09:48:50.759784 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="bea34349-b8c1-4e5b-b2bb-839fcd0eb76d" containerName="container-00" Dec 01 09:48:50 crc kubenswrapper[4873]: I1201 09:48:50.761457 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s826v" Dec 01 09:48:50 crc kubenswrapper[4873]: I1201 09:48:50.771208 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-s826v"] Dec 01 09:48:50 crc kubenswrapper[4873]: I1201 09:48:50.909950 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5l7sw\" (UniqueName: \"kubernetes.io/projected/ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7-kube-api-access-5l7sw\") pod \"redhat-marketplace-s826v\" (UID: \"ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7\") " pod="openshift-marketplace/redhat-marketplace-s826v" Dec 01 09:48:50 crc kubenswrapper[4873]: I1201 09:48:50.910031 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7-catalog-content\") pod \"redhat-marketplace-s826v\" (UID: \"ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7\") " pod="openshift-marketplace/redhat-marketplace-s826v" Dec 01 09:48:50 crc kubenswrapper[4873]: I1201 09:48:50.910158 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7-utilities\") pod \"redhat-marketplace-s826v\" (UID: \"ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7\") " pod="openshift-marketplace/redhat-marketplace-s826v" Dec 01 09:48:51 crc kubenswrapper[4873]: I1201 09:48:51.012057 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5l7sw\" (UniqueName: \"kubernetes.io/projected/ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7-kube-api-access-5l7sw\") pod \"redhat-marketplace-s826v\" (UID: \"ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7\") " pod="openshift-marketplace/redhat-marketplace-s826v" Dec 01 09:48:51 crc kubenswrapper[4873]: I1201 09:48:51.012129 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7-catalog-content\") pod \"redhat-marketplace-s826v\" (UID: \"ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7\") " pod="openshift-marketplace/redhat-marketplace-s826v" Dec 01 09:48:51 crc kubenswrapper[4873]: I1201 09:48:51.012288 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7-utilities\") pod \"redhat-marketplace-s826v\" (UID: \"ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7\") " pod="openshift-marketplace/redhat-marketplace-s826v" Dec 01 09:48:51 crc kubenswrapper[4873]: I1201 09:48:51.012912 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7-utilities\") pod \"redhat-marketplace-s826v\" (UID: \"ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7\") " pod="openshift-marketplace/redhat-marketplace-s826v" Dec 01 09:48:51 crc kubenswrapper[4873]: I1201 09:48:51.012936 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7-catalog-content\") pod \"redhat-marketplace-s826v\" (UID: \"ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7\") " pod="openshift-marketplace/redhat-marketplace-s826v" Dec 01 09:48:51 crc kubenswrapper[4873]: I1201 09:48:51.034732 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5l7sw\" (UniqueName: \"kubernetes.io/projected/ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7-kube-api-access-5l7sw\") pod \"redhat-marketplace-s826v\" (UID: \"ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7\") " pod="openshift-marketplace/redhat-marketplace-s826v" Dec 01 09:48:51 crc kubenswrapper[4873]: I1201 09:48:51.108313 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s826v" Dec 01 09:48:51 crc kubenswrapper[4873]: I1201 09:48:51.606804 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-s826v"] Dec 01 09:48:51 crc kubenswrapper[4873]: I1201 09:48:51.998239 4873 generic.go:334] "Generic (PLEG): container finished" podID="ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7" containerID="54caf495b8e14d0a8c366e27116d3fbcc1e05fc44752d6d042394223ace00d3d" exitCode=0 Dec 01 09:48:51 crc kubenswrapper[4873]: I1201 09:48:51.998334 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s826v" event={"ID":"ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7","Type":"ContainerDied","Data":"54caf495b8e14d0a8c366e27116d3fbcc1e05fc44752d6d042394223ace00d3d"} Dec 01 09:48:51 crc kubenswrapper[4873]: I1201 09:48:51.998620 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s826v" event={"ID":"ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7","Type":"ContainerStarted","Data":"a39d6005f5be0fe7a7867fb3a1864f524937b1bfcd2b27b5a46b93be67fd9114"} Dec 01 09:48:53 crc kubenswrapper[4873]: I1201 09:48:53.013524 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s826v" event={"ID":"ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7","Type":"ContainerStarted","Data":"923f89e38c1ece5e258a933b3ccb107bda5242e7cd7f2673216a32f00f56eaee"} Dec 01 09:48:54 crc kubenswrapper[4873]: I1201 09:48:54.026189 4873 generic.go:334] "Generic (PLEG): container finished" podID="ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7" containerID="923f89e38c1ece5e258a933b3ccb107bda5242e7cd7f2673216a32f00f56eaee" exitCode=0 Dec 01 09:48:54 crc kubenswrapper[4873]: I1201 09:48:54.026306 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s826v" event={"ID":"ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7","Type":"ContainerDied","Data":"923f89e38c1ece5e258a933b3ccb107bda5242e7cd7f2673216a32f00f56eaee"} Dec 01 09:48:56 crc kubenswrapper[4873]: I1201 09:48:56.087964 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s826v" event={"ID":"ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7","Type":"ContainerStarted","Data":"25c04b3a127631b5900c159d559517c4d976aad94feeebbdd83df157cc77ff30"} Dec 01 09:48:56 crc kubenswrapper[4873]: I1201 09:48:56.118890 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-s826v" podStartSLOduration=3.159144669 podStartE2EDuration="6.11886595s" podCreationTimestamp="2025-12-01 09:48:50 +0000 UTC" firstStartedPulling="2025-12-01 09:48:52.000512987 +0000 UTC m=+4107.902621536" lastFinishedPulling="2025-12-01 09:48:54.960234278 +0000 UTC m=+4110.862342817" observedRunningTime="2025-12-01 09:48:56.109963539 +0000 UTC m=+4112.012072088" watchObservedRunningTime="2025-12-01 09:48:56.11886595 +0000 UTC m=+4112.020974489" Dec 01 09:49:01 crc kubenswrapper[4873]: I1201 09:49:01.109081 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-s826v" Dec 01 09:49:01 crc kubenswrapper[4873]: I1201 09:49:01.111090 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-s826v" Dec 01 09:49:01 crc kubenswrapper[4873]: I1201 09:49:01.163665 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-s826v" Dec 01 09:49:02 crc kubenswrapper[4873]: I1201 09:49:02.196964 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-s826v" Dec 01 09:49:02 crc kubenswrapper[4873]: I1201 09:49:02.258210 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-s826v"] Dec 01 09:49:04 crc kubenswrapper[4873]: I1201 09:49:04.163044 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-s826v" podUID="ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7" containerName="registry-server" containerID="cri-o://25c04b3a127631b5900c159d559517c4d976aad94feeebbdd83df157cc77ff30" gracePeriod=2 Dec 01 09:49:04 crc kubenswrapper[4873]: I1201 09:49:04.674428 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s826v" Dec 01 09:49:04 crc kubenswrapper[4873]: I1201 09:49:04.803573 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7-utilities\") pod \"ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7\" (UID: \"ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7\") " Dec 01 09:49:04 crc kubenswrapper[4873]: I1201 09:49:04.803710 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7-catalog-content\") pod \"ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7\" (UID: \"ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7\") " Dec 01 09:49:04 crc kubenswrapper[4873]: I1201 09:49:04.803891 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5l7sw\" (UniqueName: \"kubernetes.io/projected/ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7-kube-api-access-5l7sw\") pod \"ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7\" (UID: \"ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7\") " Dec 01 09:49:04 crc kubenswrapper[4873]: I1201 09:49:04.804807 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7-utilities" (OuterVolumeSpecName: "utilities") pod "ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7" (UID: "ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:49:04 crc kubenswrapper[4873]: I1201 09:49:04.821601 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7-kube-api-access-5l7sw" (OuterVolumeSpecName: "kube-api-access-5l7sw") pod "ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7" (UID: "ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7"). InnerVolumeSpecName "kube-api-access-5l7sw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:49:04 crc kubenswrapper[4873]: I1201 09:49:04.837286 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7" (UID: "ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:49:04 crc kubenswrapper[4873]: I1201 09:49:04.907087 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5l7sw\" (UniqueName: \"kubernetes.io/projected/ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7-kube-api-access-5l7sw\") on node \"crc\" DevicePath \"\"" Dec 01 09:49:04 crc kubenswrapper[4873]: I1201 09:49:04.907439 4873 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 09:49:04 crc kubenswrapper[4873]: I1201 09:49:04.907449 4873 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 09:49:05 crc kubenswrapper[4873]: I1201 09:49:05.176815 4873 generic.go:334] "Generic (PLEG): container finished" podID="ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7" containerID="25c04b3a127631b5900c159d559517c4d976aad94feeebbdd83df157cc77ff30" exitCode=0 Dec 01 09:49:05 crc kubenswrapper[4873]: I1201 09:49:05.176912 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s826v" event={"ID":"ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7","Type":"ContainerDied","Data":"25c04b3a127631b5900c159d559517c4d976aad94feeebbdd83df157cc77ff30"} Dec 01 09:49:05 crc kubenswrapper[4873]: I1201 09:49:05.176950 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s826v" Dec 01 09:49:05 crc kubenswrapper[4873]: I1201 09:49:05.176990 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s826v" event={"ID":"ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7","Type":"ContainerDied","Data":"a39d6005f5be0fe7a7867fb3a1864f524937b1bfcd2b27b5a46b93be67fd9114"} Dec 01 09:49:05 crc kubenswrapper[4873]: I1201 09:49:05.177035 4873 scope.go:117] "RemoveContainer" containerID="25c04b3a127631b5900c159d559517c4d976aad94feeebbdd83df157cc77ff30" Dec 01 09:49:05 crc kubenswrapper[4873]: I1201 09:49:05.209628 4873 scope.go:117] "RemoveContainer" containerID="923f89e38c1ece5e258a933b3ccb107bda5242e7cd7f2673216a32f00f56eaee" Dec 01 09:49:05 crc kubenswrapper[4873]: I1201 09:49:05.216080 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-s826v"] Dec 01 09:49:05 crc kubenswrapper[4873]: I1201 09:49:05.229981 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-s826v"] Dec 01 09:49:05 crc kubenswrapper[4873]: I1201 09:49:05.245472 4873 scope.go:117] "RemoveContainer" containerID="54caf495b8e14d0a8c366e27116d3fbcc1e05fc44752d6d042394223ace00d3d" Dec 01 09:49:05 crc kubenswrapper[4873]: I1201 09:49:05.295536 4873 scope.go:117] "RemoveContainer" containerID="25c04b3a127631b5900c159d559517c4d976aad94feeebbdd83df157cc77ff30" Dec 01 09:49:05 crc kubenswrapper[4873]: E1201 09:49:05.296658 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25c04b3a127631b5900c159d559517c4d976aad94feeebbdd83df157cc77ff30\": container with ID starting with 25c04b3a127631b5900c159d559517c4d976aad94feeebbdd83df157cc77ff30 not found: ID does not exist" containerID="25c04b3a127631b5900c159d559517c4d976aad94feeebbdd83df157cc77ff30" Dec 01 09:49:05 crc kubenswrapper[4873]: I1201 09:49:05.296721 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25c04b3a127631b5900c159d559517c4d976aad94feeebbdd83df157cc77ff30"} err="failed to get container status \"25c04b3a127631b5900c159d559517c4d976aad94feeebbdd83df157cc77ff30\": rpc error: code = NotFound desc = could not find container \"25c04b3a127631b5900c159d559517c4d976aad94feeebbdd83df157cc77ff30\": container with ID starting with 25c04b3a127631b5900c159d559517c4d976aad94feeebbdd83df157cc77ff30 not found: ID does not exist" Dec 01 09:49:05 crc kubenswrapper[4873]: I1201 09:49:05.296783 4873 scope.go:117] "RemoveContainer" containerID="923f89e38c1ece5e258a933b3ccb107bda5242e7cd7f2673216a32f00f56eaee" Dec 01 09:49:05 crc kubenswrapper[4873]: E1201 09:49:05.297457 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"923f89e38c1ece5e258a933b3ccb107bda5242e7cd7f2673216a32f00f56eaee\": container with ID starting with 923f89e38c1ece5e258a933b3ccb107bda5242e7cd7f2673216a32f00f56eaee not found: ID does not exist" containerID="923f89e38c1ece5e258a933b3ccb107bda5242e7cd7f2673216a32f00f56eaee" Dec 01 09:49:05 crc kubenswrapper[4873]: I1201 09:49:05.297490 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"923f89e38c1ece5e258a933b3ccb107bda5242e7cd7f2673216a32f00f56eaee"} err="failed to get container status \"923f89e38c1ece5e258a933b3ccb107bda5242e7cd7f2673216a32f00f56eaee\": rpc error: code = NotFound desc = could not find container \"923f89e38c1ece5e258a933b3ccb107bda5242e7cd7f2673216a32f00f56eaee\": container with ID starting with 923f89e38c1ece5e258a933b3ccb107bda5242e7cd7f2673216a32f00f56eaee not found: ID does not exist" Dec 01 09:49:05 crc kubenswrapper[4873]: I1201 09:49:05.297510 4873 scope.go:117] "RemoveContainer" containerID="54caf495b8e14d0a8c366e27116d3fbcc1e05fc44752d6d042394223ace00d3d" Dec 01 09:49:05 crc kubenswrapper[4873]: E1201 09:49:05.298096 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54caf495b8e14d0a8c366e27116d3fbcc1e05fc44752d6d042394223ace00d3d\": container with ID starting with 54caf495b8e14d0a8c366e27116d3fbcc1e05fc44752d6d042394223ace00d3d not found: ID does not exist" containerID="54caf495b8e14d0a8c366e27116d3fbcc1e05fc44752d6d042394223ace00d3d" Dec 01 09:49:05 crc kubenswrapper[4873]: I1201 09:49:05.298151 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54caf495b8e14d0a8c366e27116d3fbcc1e05fc44752d6d042394223ace00d3d"} err="failed to get container status \"54caf495b8e14d0a8c366e27116d3fbcc1e05fc44752d6d042394223ace00d3d\": rpc error: code = NotFound desc = could not find container \"54caf495b8e14d0a8c366e27116d3fbcc1e05fc44752d6d042394223ace00d3d\": container with ID starting with 54caf495b8e14d0a8c366e27116d3fbcc1e05fc44752d6d042394223ace00d3d not found: ID does not exist" Dec 01 09:49:06 crc kubenswrapper[4873]: I1201 09:49:06.444764 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7" path="/var/lib/kubelet/pods/ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7/volumes" Dec 01 09:49:20 crc kubenswrapper[4873]: I1201 09:49:20.038402 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7fc6f67df6-72xpt_145a11d3-f1db-45aa-a3fe-0d4709905406/barbican-api/0.log" Dec 01 09:49:20 crc kubenswrapper[4873]: I1201 09:49:20.186574 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7fc6f67df6-72xpt_145a11d3-f1db-45aa-a3fe-0d4709905406/barbican-api-log/0.log" Dec 01 09:49:20 crc kubenswrapper[4873]: I1201 09:49:20.331856 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6674fb675b-58jdf_99f7daba-8483-4494-b783-9628f828ea49/barbican-keystone-listener/0.log" Dec 01 09:49:20 crc kubenswrapper[4873]: I1201 09:49:20.394249 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6674fb675b-58jdf_99f7daba-8483-4494-b783-9628f828ea49/barbican-keystone-listener-log/0.log" Dec 01 09:49:20 crc kubenswrapper[4873]: I1201 09:49:20.556484 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-9b48ddc7c-fn75z_4d0787b5-0da8-493b-b55d-b6e36a759f7f/barbican-worker/0.log" Dec 01 09:49:20 crc kubenswrapper[4873]: I1201 09:49:20.632003 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-9b48ddc7c-fn75z_4d0787b5-0da8-493b-b55d-b6e36a759f7f/barbican-worker-log/0.log" Dec 01 09:49:20 crc kubenswrapper[4873]: I1201 09:49:20.863871 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-wfccb_cd2a1185-cebc-4f89-88c2-63cfebe97759/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:49:20 crc kubenswrapper[4873]: I1201 09:49:20.942758 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_6ccdf302-89b3-4dc2-94c4-45ea2c2ea241/ceilometer-central-agent/0.log" Dec 01 09:49:21 crc kubenswrapper[4873]: I1201 09:49:21.062083 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_6ccdf302-89b3-4dc2-94c4-45ea2c2ea241/ceilometer-notification-agent/0.log" Dec 01 09:49:21 crc kubenswrapper[4873]: I1201 09:49:21.103749 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_6ccdf302-89b3-4dc2-94c4-45ea2c2ea241/proxy-httpd/0.log" Dec 01 09:49:21 crc kubenswrapper[4873]: I1201 09:49:21.154148 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_6ccdf302-89b3-4dc2-94c4-45ea2c2ea241/sg-core/0.log" Dec 01 09:49:21 crc kubenswrapper[4873]: I1201 09:49:21.327444 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-edpm-deployment-openstack-edpm-ipam-cz28q_d186292e-15d4-4944-b6ac-7db1fb7b85f5/ceph-client-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:49:21 crc kubenswrapper[4873]: I1201 09:49:21.366417 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-75xtr_b355b9ce-a737-407d-8c9e-5d9db024bb10/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:49:21 crc kubenswrapper[4873]: I1201 09:49:21.617598 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_50bc343c-5aca-46d3-a9de-31546ac7c45f/cinder-api/0.log" Dec 01 09:49:21 crc kubenswrapper[4873]: I1201 09:49:21.714396 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_50bc343c-5aca-46d3-a9de-31546ac7c45f/cinder-api-log/0.log" Dec 01 09:49:21 crc kubenswrapper[4873]: I1201 09:49:21.915384 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_3ba752bd-4b19-4ac4-9798-60de8b5a6f3e/probe/0.log" Dec 01 09:49:22 crc kubenswrapper[4873]: I1201 09:49:22.134059 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_4c035924-7ef3-4fc1-8652-78317f6b5c70/cinder-scheduler/0.log" Dec 01 09:49:22 crc kubenswrapper[4873]: I1201 09:49:22.224316 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_4c035924-7ef3-4fc1-8652-78317f6b5c70/probe/0.log" Dec 01 09:49:22 crc kubenswrapper[4873]: I1201 09:49:22.278223 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_3ba752bd-4b19-4ac4-9798-60de8b5a6f3e/cinder-backup/0.log" Dec 01 09:49:22 crc kubenswrapper[4873]: I1201 09:49:22.527996 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_96e55f05-de52-4914-9fc8-7634bff464a7/probe/0.log" Dec 01 09:49:22 crc kubenswrapper[4873]: I1201 09:49:22.580696 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_96e55f05-de52-4914-9fc8-7634bff464a7/cinder-volume/0.log" Dec 01 09:49:22 crc kubenswrapper[4873]: I1201 09:49:22.687487 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-lcck8_99a70cc5-d0e6-4066-9fdb-7524bf04b7ab/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:49:22 crc kubenswrapper[4873]: I1201 09:49:22.907198 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-kntht_cf256dda-e402-4a9f-bff1-fe2990f7ce72/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:49:22 crc kubenswrapper[4873]: I1201 09:49:22.936684 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-76b5fdb995-hdbhs_90639568-2248-4249-be34-140ed23a5d1d/init/0.log" Dec 01 09:49:23 crc kubenswrapper[4873]: I1201 09:49:23.160726 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-76b5fdb995-hdbhs_90639568-2248-4249-be34-140ed23a5d1d/init/0.log" Dec 01 09:49:23 crc kubenswrapper[4873]: I1201 09:49:23.175316 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_380e16b7-2a33-4591-bb99-9382512be92c/glance-httpd/0.log" Dec 01 09:49:23 crc kubenswrapper[4873]: I1201 09:49:23.226535 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-76b5fdb995-hdbhs_90639568-2248-4249-be34-140ed23a5d1d/dnsmasq-dns/0.log" Dec 01 09:49:23 crc kubenswrapper[4873]: I1201 09:49:23.701636 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_380e16b7-2a33-4591-bb99-9382512be92c/glance-log/0.log" Dec 01 09:49:23 crc kubenswrapper[4873]: I1201 09:49:23.782776 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_90a86592-ced2-4308-91fc-e9993667556a/glance-httpd/0.log" Dec 01 09:49:23 crc kubenswrapper[4873]: I1201 09:49:23.815367 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_90a86592-ced2-4308-91fc-e9993667556a/glance-log/0.log" Dec 01 09:49:24 crc kubenswrapper[4873]: I1201 09:49:24.126227 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6968fbf848-w68ch_7f8ecf06-f050-454b-8d36-e91b49847601/horizon/0.log" Dec 01 09:49:24 crc kubenswrapper[4873]: I1201 09:49:24.136695 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-dzcqj_36c9f7db-7138-48d1-9e4d-f04be409c123/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:49:24 crc kubenswrapper[4873]: I1201 09:49:24.203647 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6968fbf848-w68ch_7f8ecf06-f050-454b-8d36-e91b49847601/horizon-log/0.log" Dec 01 09:49:24 crc kubenswrapper[4873]: I1201 09:49:24.419656 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-99878_83908788-7c1f-410a-be4d-79510d6703b5/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:49:24 crc kubenswrapper[4873]: I1201 09:49:24.558499 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-58576c9895-sk9tb_0857cf38-58dd-416b-92ca-6b98800ff512/keystone-api/0.log" Dec 01 09:49:24 crc kubenswrapper[4873]: I1201 09:49:24.583149 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29409661-gls8t_cf6891b3-2f06-4218-b321-5d7fac6edb7c/keystone-cron/0.log" Dec 01 09:49:24 crc kubenswrapper[4873]: I1201 09:49:24.702621 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_84cdb02b-04a2-4456-bd2e-48d9e9a896dd/kube-state-metrics/0.log" Dec 01 09:49:24 crc kubenswrapper[4873]: I1201 09:49:24.844911 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-9l9mp_328559dd-d6ca-400a-9bb1-955781e2e1ea/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:49:25 crc kubenswrapper[4873]: I1201 09:49:25.003541 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_7744a110-0501-4606-9c09-a6442c0fefb0/manila-api-log/0.log" Dec 01 09:49:25 crc kubenswrapper[4873]: I1201 09:49:25.151134 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_7744a110-0501-4606-9c09-a6442c0fefb0/manila-api/0.log" Dec 01 09:49:25 crc kubenswrapper[4873]: I1201 09:49:25.184640 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_3f333cac-7ae8-4563-a71e-3439f5b7717d/probe/0.log" Dec 01 09:49:25 crc kubenswrapper[4873]: I1201 09:49:25.219932 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_3f333cac-7ae8-4563-a71e-3439f5b7717d/manila-scheduler/0.log" Dec 01 09:49:25 crc kubenswrapper[4873]: I1201 09:49:25.387865 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_7f96788d-8f75-490b-99b7-08096baaab34/probe/0.log" Dec 01 09:49:25 crc kubenswrapper[4873]: I1201 09:49:25.399231 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_7f96788d-8f75-490b-99b7-08096baaab34/manila-share/0.log" Dec 01 09:49:25 crc kubenswrapper[4873]: I1201 09:49:25.697892 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-f99dddf57-sr27n_7b48bbdb-ef25-4a25-a79f-d01eca3d63df/neutron-httpd/0.log" Dec 01 09:49:25 crc kubenswrapper[4873]: I1201 09:49:25.732547 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-f99dddf57-sr27n_7b48bbdb-ef25-4a25-a79f-d01eca3d63df/neutron-api/0.log" Dec 01 09:49:25 crc kubenswrapper[4873]: I1201 09:49:25.981068 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-d599b_ed4f574a-846e-41be-88ce-56d89a392a87/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:49:26 crc kubenswrapper[4873]: I1201 09:49:26.234897 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_2e6c09a7-97f9-4048-b5a9-b3df685eacbc/memcached/0.log" Dec 01 09:49:26 crc kubenswrapper[4873]: I1201 09:49:26.288656 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_11dc97a8-cb16-4992-abc5-4fe5de83bac0/nova-api-log/0.log" Dec 01 09:49:26 crc kubenswrapper[4873]: I1201 09:49:26.524851 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_e18f53d4-43ea-46f1-ae6f-5eeee3ac5b5d/nova-cell0-conductor-conductor/0.log" Dec 01 09:49:26 crc kubenswrapper[4873]: I1201 09:49:26.624367 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_11dc97a8-cb16-4992-abc5-4fe5de83bac0/nova-api-api/0.log" Dec 01 09:49:26 crc kubenswrapper[4873]: I1201 09:49:26.767264 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_425ae46c-8af5-46d8-bc7c-e7056aebdd34/nova-cell1-conductor-conductor/0.log" Dec 01 09:49:26 crc kubenswrapper[4873]: I1201 09:49:26.862066 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_ae20a830-d3d8-42fc-9fff-1aabc6ac8139/nova-cell1-novncproxy-novncproxy/0.log" Dec 01 09:49:26 crc kubenswrapper[4873]: I1201 09:49:26.922207 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ztjp_38cfd1d6-bdc0-4e9c-af0d-3fd4d3d069cf/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:49:27 crc kubenswrapper[4873]: I1201 09:49:27.091036 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_e54ecbbf-ac5c-4924-a396-549da1855e65/nova-metadata-log/0.log" Dec 01 09:49:27 crc kubenswrapper[4873]: I1201 09:49:27.345999 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_714ffca1-ee33-4f92-a55e-29e2edf84eff/mysql-bootstrap/0.log" Dec 01 09:49:27 crc kubenswrapper[4873]: I1201 09:49:27.384742 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_634610e2-dd0b-46f5-a934-e7f44249c731/nova-scheduler-scheduler/0.log" Dec 01 09:49:27 crc kubenswrapper[4873]: I1201 09:49:27.569281 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_714ffca1-ee33-4f92-a55e-29e2edf84eff/mysql-bootstrap/0.log" Dec 01 09:49:27 crc kubenswrapper[4873]: I1201 09:49:27.580733 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_714ffca1-ee33-4f92-a55e-29e2edf84eff/galera/0.log" Dec 01 09:49:28 crc kubenswrapper[4873]: I1201 09:49:28.047730 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_250cbd69-c69a-475c-8973-bc98caf4b264/mysql-bootstrap/0.log" Dec 01 09:49:28 crc kubenswrapper[4873]: I1201 09:49:28.094149 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_e54ecbbf-ac5c-4924-a396-549da1855e65/nova-metadata-metadata/0.log" Dec 01 09:49:28 crc kubenswrapper[4873]: I1201 09:49:28.223544 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_250cbd69-c69a-475c-8973-bc98caf4b264/mysql-bootstrap/0.log" Dec 01 09:49:28 crc kubenswrapper[4873]: I1201 09:49:28.255551 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_250cbd69-c69a-475c-8973-bc98caf4b264/galera/0.log" Dec 01 09:49:28 crc kubenswrapper[4873]: I1201 09:49:28.298350 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_43a97c60-b43e-4896-9592-d41863b1c203/openstackclient/0.log" Dec 01 09:49:28 crc kubenswrapper[4873]: I1201 09:49:28.522289 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-7xdqf_93fb09e4-51b3-44f7-afab-c49c374c7f56/openstack-network-exporter/0.log" Dec 01 09:49:28 crc kubenswrapper[4873]: I1201 09:49:28.545342 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-xd66n_243fec55-bb19-4807-8b15-69584a922cf6/ovsdb-server-init/0.log" Dec 01 09:49:28 crc kubenswrapper[4873]: I1201 09:49:28.723408 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-xd66n_243fec55-bb19-4807-8b15-69584a922cf6/ovsdb-server-init/0.log" Dec 01 09:49:28 crc kubenswrapper[4873]: I1201 09:49:28.743083 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-xd66n_243fec55-bb19-4807-8b15-69584a922cf6/ovs-vswitchd/0.log" Dec 01 09:49:28 crc kubenswrapper[4873]: I1201 09:49:28.752407 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-w8v6j_b79364c5-00b2-4ab8-b2f0-aff17ed902b4/ovn-controller/0.log" Dec 01 09:49:28 crc kubenswrapper[4873]: I1201 09:49:28.786035 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-xd66n_243fec55-bb19-4807-8b15-69584a922cf6/ovsdb-server/0.log" Dec 01 09:49:28 crc kubenswrapper[4873]: I1201 09:49:28.971098 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-q7cl2_0b014b12-82de-4f51-84bc-c8d41241f672/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:49:29 crc kubenswrapper[4873]: I1201 09:49:29.011417 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_a5cf24b4-dd05-45a4-bf8f-5895c30ada04/openstack-network-exporter/0.log" Dec 01 09:49:29 crc kubenswrapper[4873]: I1201 09:49:29.048236 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_a5cf24b4-dd05-45a4-bf8f-5895c30ada04/ovn-northd/0.log" Dec 01 09:49:29 crc kubenswrapper[4873]: I1201 09:49:29.217032 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_21c27fff-80e6-4a74-aad2-5e772b811535/openstack-network-exporter/0.log" Dec 01 09:49:29 crc kubenswrapper[4873]: I1201 09:49:29.233237 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_21c27fff-80e6-4a74-aad2-5e772b811535/ovsdbserver-nb/0.log" Dec 01 09:49:29 crc kubenswrapper[4873]: I1201 09:49:29.319326 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_dbe0e07d-50a3-47c2-831d-69f24c0e838a/openstack-network-exporter/0.log" Dec 01 09:49:29 crc kubenswrapper[4873]: I1201 09:49:29.468625 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7d689bd78d-s9rtj_6615f806-2d35-4500-a746-3fab7fbfde88/placement-api/0.log" Dec 01 09:49:29 crc kubenswrapper[4873]: I1201 09:49:29.475740 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_dbe0e07d-50a3-47c2-831d-69f24c0e838a/ovsdbserver-sb/0.log" Dec 01 09:49:29 crc kubenswrapper[4873]: I1201 09:49:29.633329 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7d689bd78d-s9rtj_6615f806-2d35-4500-a746-3fab7fbfde88/placement-log/0.log" Dec 01 09:49:29 crc kubenswrapper[4873]: I1201 09:49:29.710840 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3baa8150-922b-4de7-ae44-24b9b5584212/setup-container/0.log" Dec 01 09:49:29 crc kubenswrapper[4873]: I1201 09:49:29.923554 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3baa8150-922b-4de7-ae44-24b9b5584212/setup-container/0.log" Dec 01 09:49:29 crc kubenswrapper[4873]: I1201 09:49:29.964194 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_84aac4b8-9ed2-40d2-81f1-eba3af6e46b0/setup-container/0.log" Dec 01 09:49:29 crc kubenswrapper[4873]: I1201 09:49:29.969054 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3baa8150-922b-4de7-ae44-24b9b5584212/rabbitmq/0.log" Dec 01 09:49:30 crc kubenswrapper[4873]: I1201 09:49:30.298638 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_84aac4b8-9ed2-40d2-81f1-eba3af6e46b0/setup-container/0.log" Dec 01 09:49:30 crc kubenswrapper[4873]: I1201 09:49:30.339719 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-9bdgf_e301f3ff-9983-4440-ad43-96ef975437f7/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:49:30 crc kubenswrapper[4873]: I1201 09:49:30.413051 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_84aac4b8-9ed2-40d2-81f1-eba3af6e46b0/rabbitmq/0.log" Dec 01 09:49:30 crc kubenswrapper[4873]: I1201 09:49:30.732901 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-g5ghr_47baf114-fcb4-46ba-ac64-53f969ebcbbc/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:49:30 crc kubenswrapper[4873]: I1201 09:49:30.943725 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-gg2dk_74115943-1cf4-48ce-9f8c-41de87f8fe52/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:49:31 crc kubenswrapper[4873]: I1201 09:49:31.067928 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-9k6m8_2ea95238-7e66-46b0-9c20-b6088f9d6737/ssh-known-hosts-edpm-deployment/0.log" Dec 01 09:49:31 crc kubenswrapper[4873]: I1201 09:49:31.244633 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_cd15a962-a485-48ad-b0af-d7a3ec0e4ab1/tempest-tests-tempest-tests-runner/0.log" Dec 01 09:49:31 crc kubenswrapper[4873]: I1201 09:49:31.334262 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_f210543d-a2a6-4746-b279-04fe37ee1b76/test-operator-logs-container/0.log" Dec 01 09:49:31 crc kubenswrapper[4873]: I1201 09:49:31.488894 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-qt48t_933df0ea-eae8-49a3-affc-2f421e3ae777/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:49:58 crc kubenswrapper[4873]: I1201 09:49:58.463925 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-5rq8z_51253ba6-e0b5-44ac-8c18-be17b4d13024/kube-rbac-proxy/0.log" Dec 01 09:49:58 crc kubenswrapper[4873]: I1201 09:49:58.532280 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-5rq8z_51253ba6-e0b5-44ac-8c18-be17b4d13024/manager/0.log" Dec 01 09:49:58 crc kubenswrapper[4873]: I1201 09:49:58.691563 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-gqm5w_671ad3cc-9716-42de-b5ae-fb69847d3bd7/kube-rbac-proxy/0.log" Dec 01 09:49:58 crc kubenswrapper[4873]: I1201 09:49:58.723352 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-gqm5w_671ad3cc-9716-42de-b5ae-fb69847d3bd7/manager/0.log" Dec 01 09:49:58 crc kubenswrapper[4873]: I1201 09:49:58.868467 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d92c7d608fded6ff7325ca25db46cfa3dc2f9a2773cb154702b56fef16ks4c9_08dfa43f-80d7-4785-a0ce-c8b6d8d95351/util/0.log" Dec 01 09:49:59 crc kubenswrapper[4873]: I1201 09:49:59.049513 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d92c7d608fded6ff7325ca25db46cfa3dc2f9a2773cb154702b56fef16ks4c9_08dfa43f-80d7-4785-a0ce-c8b6d8d95351/util/0.log" Dec 01 09:49:59 crc kubenswrapper[4873]: I1201 09:49:59.074463 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d92c7d608fded6ff7325ca25db46cfa3dc2f9a2773cb154702b56fef16ks4c9_08dfa43f-80d7-4785-a0ce-c8b6d8d95351/pull/0.log" Dec 01 09:49:59 crc kubenswrapper[4873]: I1201 09:49:59.093524 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d92c7d608fded6ff7325ca25db46cfa3dc2f9a2773cb154702b56fef16ks4c9_08dfa43f-80d7-4785-a0ce-c8b6d8d95351/pull/0.log" Dec 01 09:49:59 crc kubenswrapper[4873]: I1201 09:49:59.253746 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d92c7d608fded6ff7325ca25db46cfa3dc2f9a2773cb154702b56fef16ks4c9_08dfa43f-80d7-4785-a0ce-c8b6d8d95351/util/0.log" Dec 01 09:49:59 crc kubenswrapper[4873]: I1201 09:49:59.277942 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d92c7d608fded6ff7325ca25db46cfa3dc2f9a2773cb154702b56fef16ks4c9_08dfa43f-80d7-4785-a0ce-c8b6d8d95351/pull/0.log" Dec 01 09:49:59 crc kubenswrapper[4873]: I1201 09:49:59.295687 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d92c7d608fded6ff7325ca25db46cfa3dc2f9a2773cb154702b56fef16ks4c9_08dfa43f-80d7-4785-a0ce-c8b6d8d95351/extract/0.log" Dec 01 09:49:59 crc kubenswrapper[4873]: I1201 09:49:59.491949 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-6r8hb_723f446d-21bb-432a-b2af-93219593819d/manager/0.log" Dec 01 09:49:59 crc kubenswrapper[4873]: I1201 09:49:59.512439 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-6r8hb_723f446d-21bb-432a-b2af-93219593819d/kube-rbac-proxy/0.log" Dec 01 09:49:59 crc kubenswrapper[4873]: I1201 09:49:59.570825 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5cd474b4ff-wcmv4_f778fd6c-4c8a-4067-b3f2-cb7d98a50bf7/kube-rbac-proxy/0.log" Dec 01 09:49:59 crc kubenswrapper[4873]: I1201 09:49:59.778273 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5cd474b4ff-wcmv4_f778fd6c-4c8a-4067-b3f2-cb7d98a50bf7/manager/0.log" Dec 01 09:49:59 crc kubenswrapper[4873]: I1201 09:49:59.794688 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-d7c6w_967215e8-7a18-4507-96c1-4c79c7e1d51a/manager/0.log" Dec 01 09:49:59 crc kubenswrapper[4873]: I1201 09:49:59.905544 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-d7c6w_967215e8-7a18-4507-96c1-4c79c7e1d51a/kube-rbac-proxy/0.log" Dec 01 09:50:00 crc kubenswrapper[4873]: I1201 09:50:00.102250 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-45whz_94e7345b-b1ac-46e2-be25-9d64d3d33523/kube-rbac-proxy/0.log" Dec 01 09:50:00 crc kubenswrapper[4873]: I1201 09:50:00.107628 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-45whz_94e7345b-b1ac-46e2-be25-9d64d3d33523/manager/0.log" Dec 01 09:50:00 crc kubenswrapper[4873]: I1201 09:50:00.346435 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-tmdj7_4dcc939e-b548-4fb6-814a-30e2aaa8a94a/kube-rbac-proxy/0.log" Dec 01 09:50:00 crc kubenswrapper[4873]: I1201 09:50:00.410549 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-7wmqg_eb53d542-63e9-487d-9d06-237c4b2b9252/kube-rbac-proxy/0.log" Dec 01 09:50:00 crc kubenswrapper[4873]: I1201 09:50:00.508054 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-tmdj7_4dcc939e-b548-4fb6-814a-30e2aaa8a94a/manager/0.log" Dec 01 09:50:00 crc kubenswrapper[4873]: I1201 09:50:00.571656 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-7wmqg_eb53d542-63e9-487d-9d06-237c4b2b9252/manager/0.log" Dec 01 09:50:00 crc kubenswrapper[4873]: I1201 09:50:00.685700 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-546d4bdf48-pfphw_035095bd-1b65-4895-95a1-59feee524920/kube-rbac-proxy/0.log" Dec 01 09:50:00 crc kubenswrapper[4873]: I1201 09:50:00.778579 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-546d4bdf48-pfphw_035095bd-1b65-4895-95a1-59feee524920/manager/0.log" Dec 01 09:50:00 crc kubenswrapper[4873]: I1201 09:50:00.883559 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6546668bfd-j2vnt_b45909ce-2ea4-4cf2-9351-c8839e44d734/kube-rbac-proxy/0.log" Dec 01 09:50:01 crc kubenswrapper[4873]: I1201 09:50:01.003219 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6546668bfd-j2vnt_b45909ce-2ea4-4cf2-9351-c8839e44d734/manager/0.log" Dec 01 09:50:01 crc kubenswrapper[4873]: I1201 09:50:01.065209 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-hnzzf_3eff7ce2-9ff7-413c-b472-9e114e7130ca/kube-rbac-proxy/0.log" Dec 01 09:50:01 crc kubenswrapper[4873]: I1201 09:50:01.188819 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-hnzzf_3eff7ce2-9ff7-413c-b472-9e114e7130ca/manager/0.log" Dec 01 09:50:01 crc kubenswrapper[4873]: I1201 09:50:01.288340 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-sjw48_5ec1188a-c7f9-4cc8-8ea9-c1f1977041a5/kube-rbac-proxy/0.log" Dec 01 09:50:01 crc kubenswrapper[4873]: I1201 09:50:01.432970 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-sjw48_5ec1188a-c7f9-4cc8-8ea9-c1f1977041a5/manager/0.log" Dec 01 09:50:01 crc kubenswrapper[4873]: I1201 09:50:01.537206 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-x9kc4_450a3417-0074-4223-b1e3-aa1b854320fe/kube-rbac-proxy/0.log" Dec 01 09:50:01 crc kubenswrapper[4873]: I1201 09:50:01.576394 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-x9kc4_450a3417-0074-4223-b1e3-aa1b854320fe/manager/0.log" Dec 01 09:50:01 crc kubenswrapper[4873]: I1201 09:50:01.689704 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-n79q9_7e3a2af8-3381-46e2-8c23-41aab8fd1a5e/kube-rbac-proxy/0.log" Dec 01 09:50:01 crc kubenswrapper[4873]: I1201 09:50:01.773684 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-n79q9_7e3a2af8-3381-46e2-8c23-41aab8fd1a5e/manager/0.log" Dec 01 09:50:01 crc kubenswrapper[4873]: I1201 09:50:01.901181 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4vf8qt_5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5/kube-rbac-proxy/0.log" Dec 01 09:50:01 crc kubenswrapper[4873]: I1201 09:50:01.948593 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4vf8qt_5b16e7ed-6a38-4ed6-85f8-b3c2e7cfcbb5/manager/0.log" Dec 01 09:50:02 crc kubenswrapper[4873]: I1201 09:50:02.444720 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-mqm6p_b1309d4c-45e2-444c-afc5-56aa75b9abf5/registry-server/0.log" Dec 01 09:50:02 crc kubenswrapper[4873]: I1201 09:50:02.543788 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-7f57cdcc9-jttc7_c25dac67-107f-43e2-a63e-5843ff31abc8/operator/0.log" Dec 01 09:50:02 crc kubenswrapper[4873]: I1201 09:50:02.595334 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-49bz9_879e02ea-306e-4e7b-9012-cb1cb6bdee00/kube-rbac-proxy/0.log" Dec 01 09:50:02 crc kubenswrapper[4873]: I1201 09:50:02.801997 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-49bz9_879e02ea-306e-4e7b-9012-cb1cb6bdee00/manager/0.log" Dec 01 09:50:02 crc kubenswrapper[4873]: I1201 09:50:02.808916 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-v7bvz_3e0ad069-d971-4fd0-93df-b6a8794afe00/kube-rbac-proxy/0.log" Dec 01 09:50:02 crc kubenswrapper[4873]: I1201 09:50:02.891098 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-v7bvz_3e0ad069-d971-4fd0-93df-b6a8794afe00/manager/0.log" Dec 01 09:50:03 crc kubenswrapper[4873]: I1201 09:50:03.132861 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-6dxcw_f7a7bcca-9403-4e2d-bc28-53d4eb5ae252/kube-rbac-proxy/0.log" Dec 01 09:50:03 crc kubenswrapper[4873]: I1201 09:50:03.178226 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-cb9qn_664e20ef-b15c-4903-b72d-e18c7077e888/operator/0.log" Dec 01 09:50:03 crc kubenswrapper[4873]: I1201 09:50:03.312403 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-6dxcw_f7a7bcca-9403-4e2d-bc28-53d4eb5ae252/manager/0.log" Dec 01 09:50:03 crc kubenswrapper[4873]: I1201 09:50:03.659213 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7f4d6cf897-fddkc_223a64ee-92cb-4b29-91bf-ffa7ed7d64ce/manager/0.log" Dec 01 09:50:03 crc kubenswrapper[4873]: I1201 09:50:03.948235 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-bhrpn_225323c8-ed2f-4573-920d-f43f3e4561de/kube-rbac-proxy/0.log" Dec 01 09:50:03 crc kubenswrapper[4873]: I1201 09:50:03.956296 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-mvc26_34e3726b-c9f5-4a86-9e87-e179f8dae739/kube-rbac-proxy/0.log" Dec 01 09:50:04 crc kubenswrapper[4873]: I1201 09:50:04.064602 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-bhrpn_225323c8-ed2f-4573-920d-f43f3e4561de/manager/0.log" Dec 01 09:50:04 crc kubenswrapper[4873]: I1201 09:50:04.172415 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-mvc26_34e3726b-c9f5-4a86-9e87-e179f8dae739/manager/0.log" Dec 01 09:50:04 crc kubenswrapper[4873]: I1201 09:50:04.263945 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-dx4jm_1cd3aa85-6333-4cfd-aacf-e51169ef4b42/kube-rbac-proxy/0.log" Dec 01 09:50:04 crc kubenswrapper[4873]: I1201 09:50:04.285720 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-dx4jm_1cd3aa85-6333-4cfd-aacf-e51169ef4b42/manager/0.log" Dec 01 09:50:25 crc kubenswrapper[4873]: I1201 09:50:25.660726 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-xg89d_f4709124-2f04-4d3c-abb7-30a5d0925b6c/control-plane-machine-set-operator/0.log" Dec 01 09:50:25 crc kubenswrapper[4873]: I1201 09:50:25.908420 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-52pjn_fb8fcfca-0933-4a74-85de-859fda0153d6/machine-api-operator/0.log" Dec 01 09:50:25 crc kubenswrapper[4873]: I1201 09:50:25.918426 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-52pjn_fb8fcfca-0933-4a74-85de-859fda0153d6/kube-rbac-proxy/0.log" Dec 01 09:50:31 crc kubenswrapper[4873]: I1201 09:50:31.059224 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:50:31 crc kubenswrapper[4873]: I1201 09:50:31.061030 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:50:41 crc kubenswrapper[4873]: I1201 09:50:41.136534 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-blv72_c52cc19c-1588-49cd-a54b-072c33505a5a/cert-manager-controller/0.log" Dec 01 09:50:42 crc kubenswrapper[4873]: I1201 09:50:42.015714 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-47tnm_3876e535-8b83-41bf-8b75-2b21a25e1ff4/cert-manager-cainjector/0.log" Dec 01 09:50:42 crc kubenswrapper[4873]: I1201 09:50:42.142923 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-tj5cg_eb0ae03d-2543-42c9-9c93-628cab31efe2/cert-manager-webhook/0.log" Dec 01 09:50:56 crc kubenswrapper[4873]: I1201 09:50:56.544076 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-rtjft_244acac6-afd2-4d7b-b1a0-085a24cbf1c8/nmstate-console-plugin/0.log" Dec 01 09:50:56 crc kubenswrapper[4873]: I1201 09:50:56.774631 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-x8slw_5ca9fe36-a6a0-402c-a744-4125d0cd37c6/nmstate-handler/0.log" Dec 01 09:50:56 crc kubenswrapper[4873]: I1201 09:50:56.778827 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-m7d79_1c32a501-0496-45b2-8e04-984bccb3c03d/kube-rbac-proxy/0.log" Dec 01 09:50:56 crc kubenswrapper[4873]: I1201 09:50:56.836225 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-m7d79_1c32a501-0496-45b2-8e04-984bccb3c03d/nmstate-metrics/0.log" Dec 01 09:50:57 crc kubenswrapper[4873]: I1201 09:50:57.031746 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-fbds9_a61c7c72-a083-41c2-b0a4-707d26b095c6/nmstate-operator/0.log" Dec 01 09:50:57 crc kubenswrapper[4873]: I1201 09:50:57.089678 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-79fqx_6050180f-b828-43a2-911c-4a44354d4bf6/nmstate-webhook/0.log" Dec 01 09:51:01 crc kubenswrapper[4873]: I1201 09:51:01.059483 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:51:01 crc kubenswrapper[4873]: I1201 09:51:01.060172 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:51:06 crc kubenswrapper[4873]: I1201 09:51:06.493199 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kj4zj"] Dec 01 09:51:06 crc kubenswrapper[4873]: E1201 09:51:06.494833 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7" containerName="registry-server" Dec 01 09:51:06 crc kubenswrapper[4873]: I1201 09:51:06.494851 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7" containerName="registry-server" Dec 01 09:51:06 crc kubenswrapper[4873]: E1201 09:51:06.494861 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7" containerName="extract-utilities" Dec 01 09:51:06 crc kubenswrapper[4873]: I1201 09:51:06.494871 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7" containerName="extract-utilities" Dec 01 09:51:06 crc kubenswrapper[4873]: E1201 09:51:06.494884 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7" containerName="extract-content" Dec 01 09:51:06 crc kubenswrapper[4873]: I1201 09:51:06.494892 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7" containerName="extract-content" Dec 01 09:51:06 crc kubenswrapper[4873]: I1201 09:51:06.495129 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee3cfe91-3bb0-4928-a5fc-58557c1a6ae7" containerName="registry-server" Dec 01 09:51:06 crc kubenswrapper[4873]: I1201 09:51:06.496897 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kj4zj" Dec 01 09:51:06 crc kubenswrapper[4873]: I1201 09:51:06.510432 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kj4zj"] Dec 01 09:51:06 crc kubenswrapper[4873]: I1201 09:51:06.644977 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90267e5e-ba3a-46ed-90b7-aba5505cc38a-utilities\") pod \"certified-operators-kj4zj\" (UID: \"90267e5e-ba3a-46ed-90b7-aba5505cc38a\") " pod="openshift-marketplace/certified-operators-kj4zj" Dec 01 09:51:06 crc kubenswrapper[4873]: I1201 09:51:06.645093 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phk2p\" (UniqueName: \"kubernetes.io/projected/90267e5e-ba3a-46ed-90b7-aba5505cc38a-kube-api-access-phk2p\") pod \"certified-operators-kj4zj\" (UID: \"90267e5e-ba3a-46ed-90b7-aba5505cc38a\") " pod="openshift-marketplace/certified-operators-kj4zj" Dec 01 09:51:06 crc kubenswrapper[4873]: I1201 09:51:06.645154 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90267e5e-ba3a-46ed-90b7-aba5505cc38a-catalog-content\") pod \"certified-operators-kj4zj\" (UID: \"90267e5e-ba3a-46ed-90b7-aba5505cc38a\") " pod="openshift-marketplace/certified-operators-kj4zj" Dec 01 09:51:06 crc kubenswrapper[4873]: I1201 09:51:06.747349 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90267e5e-ba3a-46ed-90b7-aba5505cc38a-utilities\") pod \"certified-operators-kj4zj\" (UID: \"90267e5e-ba3a-46ed-90b7-aba5505cc38a\") " pod="openshift-marketplace/certified-operators-kj4zj" Dec 01 09:51:06 crc kubenswrapper[4873]: I1201 09:51:06.747433 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phk2p\" (UniqueName: \"kubernetes.io/projected/90267e5e-ba3a-46ed-90b7-aba5505cc38a-kube-api-access-phk2p\") pod \"certified-operators-kj4zj\" (UID: \"90267e5e-ba3a-46ed-90b7-aba5505cc38a\") " pod="openshift-marketplace/certified-operators-kj4zj" Dec 01 09:51:06 crc kubenswrapper[4873]: I1201 09:51:06.747459 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90267e5e-ba3a-46ed-90b7-aba5505cc38a-catalog-content\") pod \"certified-operators-kj4zj\" (UID: \"90267e5e-ba3a-46ed-90b7-aba5505cc38a\") " pod="openshift-marketplace/certified-operators-kj4zj" Dec 01 09:51:06 crc kubenswrapper[4873]: I1201 09:51:06.748401 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90267e5e-ba3a-46ed-90b7-aba5505cc38a-utilities\") pod \"certified-operators-kj4zj\" (UID: \"90267e5e-ba3a-46ed-90b7-aba5505cc38a\") " pod="openshift-marketplace/certified-operators-kj4zj" Dec 01 09:51:06 crc kubenswrapper[4873]: I1201 09:51:06.748608 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90267e5e-ba3a-46ed-90b7-aba5505cc38a-catalog-content\") pod \"certified-operators-kj4zj\" (UID: \"90267e5e-ba3a-46ed-90b7-aba5505cc38a\") " pod="openshift-marketplace/certified-operators-kj4zj" Dec 01 09:51:06 crc kubenswrapper[4873]: I1201 09:51:06.954135 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phk2p\" (UniqueName: \"kubernetes.io/projected/90267e5e-ba3a-46ed-90b7-aba5505cc38a-kube-api-access-phk2p\") pod \"certified-operators-kj4zj\" (UID: \"90267e5e-ba3a-46ed-90b7-aba5505cc38a\") " pod="openshift-marketplace/certified-operators-kj4zj" Dec 01 09:51:07 crc kubenswrapper[4873]: I1201 09:51:07.129425 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kj4zj" Dec 01 09:51:07 crc kubenswrapper[4873]: I1201 09:51:07.654344 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kj4zj"] Dec 01 09:51:08 crc kubenswrapper[4873]: I1201 09:51:08.460833 4873 generic.go:334] "Generic (PLEG): container finished" podID="90267e5e-ba3a-46ed-90b7-aba5505cc38a" containerID="88edda7a12b2b2e25cbae08427faf9603f918d08d722e7f3f097bdcb12bc53ad" exitCode=0 Dec 01 09:51:08 crc kubenswrapper[4873]: I1201 09:51:08.460956 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kj4zj" event={"ID":"90267e5e-ba3a-46ed-90b7-aba5505cc38a","Type":"ContainerDied","Data":"88edda7a12b2b2e25cbae08427faf9603f918d08d722e7f3f097bdcb12bc53ad"} Dec 01 09:51:08 crc kubenswrapper[4873]: I1201 09:51:08.461485 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kj4zj" event={"ID":"90267e5e-ba3a-46ed-90b7-aba5505cc38a","Type":"ContainerStarted","Data":"7799074176e8722a28f9bc4f1273208a3d789b5683eded32f76655952623fddf"} Dec 01 09:51:10 crc kubenswrapper[4873]: I1201 09:51:10.528411 4873 generic.go:334] "Generic (PLEG): container finished" podID="90267e5e-ba3a-46ed-90b7-aba5505cc38a" containerID="6eb2b36c8f3d61d24ac4b0ca7e1301532f406424c8323865fec4f4f245e7227b" exitCode=0 Dec 01 09:51:10 crc kubenswrapper[4873]: I1201 09:51:10.529155 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kj4zj" event={"ID":"90267e5e-ba3a-46ed-90b7-aba5505cc38a","Type":"ContainerDied","Data":"6eb2b36c8f3d61d24ac4b0ca7e1301532f406424c8323865fec4f4f245e7227b"} Dec 01 09:51:11 crc kubenswrapper[4873]: I1201 09:51:11.544003 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kj4zj" event={"ID":"90267e5e-ba3a-46ed-90b7-aba5505cc38a","Type":"ContainerStarted","Data":"c1802b08f8bafded322f30e650309fe3f61b90a8ead290d0230b966c0dd24ae5"} Dec 01 09:51:11 crc kubenswrapper[4873]: I1201 09:51:11.571361 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kj4zj" podStartSLOduration=2.979234127 podStartE2EDuration="5.571328122s" podCreationTimestamp="2025-12-01 09:51:06 +0000 UTC" firstStartedPulling="2025-12-01 09:51:08.465112938 +0000 UTC m=+4244.367221487" lastFinishedPulling="2025-12-01 09:51:11.057206943 +0000 UTC m=+4246.959315482" observedRunningTime="2025-12-01 09:51:11.565178235 +0000 UTC m=+4247.467286794" watchObservedRunningTime="2025-12-01 09:51:11.571328122 +0000 UTC m=+4247.473436661" Dec 01 09:51:14 crc kubenswrapper[4873]: I1201 09:51:14.612580 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-6wb84_0bd118a2-14e2-438a-bea0-2cd777f71e5c/controller/0.log" Dec 01 09:51:14 crc kubenswrapper[4873]: I1201 09:51:14.786878 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-6wb84_0bd118a2-14e2-438a-bea0-2cd777f71e5c/kube-rbac-proxy/0.log" Dec 01 09:51:15 crc kubenswrapper[4873]: I1201 09:51:15.119271 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v584n_065b9940-14f0-45e2-9d92-3e7173fce9b2/cp-frr-files/0.log" Dec 01 09:51:15 crc kubenswrapper[4873]: I1201 09:51:15.320699 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v584n_065b9940-14f0-45e2-9d92-3e7173fce9b2/cp-metrics/0.log" Dec 01 09:51:15 crc kubenswrapper[4873]: I1201 09:51:15.336926 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v584n_065b9940-14f0-45e2-9d92-3e7173fce9b2/cp-frr-files/0.log" Dec 01 09:51:15 crc kubenswrapper[4873]: I1201 09:51:15.352371 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v584n_065b9940-14f0-45e2-9d92-3e7173fce9b2/cp-reloader/0.log" Dec 01 09:51:15 crc kubenswrapper[4873]: I1201 09:51:15.376386 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v584n_065b9940-14f0-45e2-9d92-3e7173fce9b2/cp-reloader/0.log" Dec 01 09:51:15 crc kubenswrapper[4873]: I1201 09:51:15.568745 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v584n_065b9940-14f0-45e2-9d92-3e7173fce9b2/cp-reloader/0.log" Dec 01 09:51:15 crc kubenswrapper[4873]: I1201 09:51:15.581003 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v584n_065b9940-14f0-45e2-9d92-3e7173fce9b2/cp-frr-files/0.log" Dec 01 09:51:15 crc kubenswrapper[4873]: I1201 09:51:15.666363 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v584n_065b9940-14f0-45e2-9d92-3e7173fce9b2/cp-metrics/0.log" Dec 01 09:51:15 crc kubenswrapper[4873]: I1201 09:51:15.669284 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v584n_065b9940-14f0-45e2-9d92-3e7173fce9b2/cp-metrics/0.log" Dec 01 09:51:15 crc kubenswrapper[4873]: I1201 09:51:15.871997 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v584n_065b9940-14f0-45e2-9d92-3e7173fce9b2/cp-frr-files/0.log" Dec 01 09:51:15 crc kubenswrapper[4873]: I1201 09:51:15.872083 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v584n_065b9940-14f0-45e2-9d92-3e7173fce9b2/cp-metrics/0.log" Dec 01 09:51:15 crc kubenswrapper[4873]: I1201 09:51:15.900776 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v584n_065b9940-14f0-45e2-9d92-3e7173fce9b2/controller/0.log" Dec 01 09:51:15 crc kubenswrapper[4873]: I1201 09:51:15.991090 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v584n_065b9940-14f0-45e2-9d92-3e7173fce9b2/cp-reloader/0.log" Dec 01 09:51:16 crc kubenswrapper[4873]: I1201 09:51:16.081152 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v584n_065b9940-14f0-45e2-9d92-3e7173fce9b2/frr-metrics/0.log" Dec 01 09:51:16 crc kubenswrapper[4873]: I1201 09:51:16.172588 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v584n_065b9940-14f0-45e2-9d92-3e7173fce9b2/kube-rbac-proxy/0.log" Dec 01 09:51:16 crc kubenswrapper[4873]: I1201 09:51:16.300185 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v584n_065b9940-14f0-45e2-9d92-3e7173fce9b2/kube-rbac-proxy-frr/0.log" Dec 01 09:51:16 crc kubenswrapper[4873]: I1201 09:51:16.471860 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v584n_065b9940-14f0-45e2-9d92-3e7173fce9b2/reloader/0.log" Dec 01 09:51:16 crc kubenswrapper[4873]: I1201 09:51:16.586803 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-ffvzp_31ad102b-37f9-491d-b017-20c6a3bc4973/frr-k8s-webhook-server/0.log" Dec 01 09:51:16 crc kubenswrapper[4873]: I1201 09:51:16.762810 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-586dc797b5-msvs2_5b1a0fb8-99da-4202-8c8e-6f5fbcec1908/manager/0.log" Dec 01 09:51:16 crc kubenswrapper[4873]: I1201 09:51:16.995225 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-767c58684d-z249t_c9747d5c-eed2-4d46-8cbf-0a01c26624da/webhook-server/0.log" Dec 01 09:51:17 crc kubenswrapper[4873]: I1201 09:51:17.109281 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-cbkfq_8f181c2f-7d70-4dd9-bcab-ae125dfa4037/kube-rbac-proxy/0.log" Dec 01 09:51:17 crc kubenswrapper[4873]: I1201 09:51:17.130783 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kj4zj" Dec 01 09:51:17 crc kubenswrapper[4873]: I1201 09:51:17.130846 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kj4zj" Dec 01 09:51:17 crc kubenswrapper[4873]: I1201 09:51:17.186853 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kj4zj" Dec 01 09:51:17 crc kubenswrapper[4873]: I1201 09:51:17.664145 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kj4zj" Dec 01 09:51:17 crc kubenswrapper[4873]: I1201 09:51:17.712966 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-cbkfq_8f181c2f-7d70-4dd9-bcab-ae125dfa4037/speaker/0.log" Dec 01 09:51:17 crc kubenswrapper[4873]: I1201 09:51:17.730652 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kj4zj"] Dec 01 09:51:17 crc kubenswrapper[4873]: I1201 09:51:17.870934 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v584n_065b9940-14f0-45e2-9d92-3e7173fce9b2/frr/0.log" Dec 01 09:51:19 crc kubenswrapper[4873]: I1201 09:51:19.669214 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kj4zj" podUID="90267e5e-ba3a-46ed-90b7-aba5505cc38a" containerName="registry-server" containerID="cri-o://c1802b08f8bafded322f30e650309fe3f61b90a8ead290d0230b966c0dd24ae5" gracePeriod=2 Dec 01 09:51:20 crc kubenswrapper[4873]: I1201 09:51:20.288983 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kj4zj" Dec 01 09:51:20 crc kubenswrapper[4873]: I1201 09:51:20.484789 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90267e5e-ba3a-46ed-90b7-aba5505cc38a-catalog-content\") pod \"90267e5e-ba3a-46ed-90b7-aba5505cc38a\" (UID: \"90267e5e-ba3a-46ed-90b7-aba5505cc38a\") " Dec 01 09:51:20 crc kubenswrapper[4873]: I1201 09:51:20.484902 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90267e5e-ba3a-46ed-90b7-aba5505cc38a-utilities\") pod \"90267e5e-ba3a-46ed-90b7-aba5505cc38a\" (UID: \"90267e5e-ba3a-46ed-90b7-aba5505cc38a\") " Dec 01 09:51:20 crc kubenswrapper[4873]: I1201 09:51:20.485116 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-phk2p\" (UniqueName: \"kubernetes.io/projected/90267e5e-ba3a-46ed-90b7-aba5505cc38a-kube-api-access-phk2p\") pod \"90267e5e-ba3a-46ed-90b7-aba5505cc38a\" (UID: \"90267e5e-ba3a-46ed-90b7-aba5505cc38a\") " Dec 01 09:51:20 crc kubenswrapper[4873]: I1201 09:51:20.487125 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90267e5e-ba3a-46ed-90b7-aba5505cc38a-utilities" (OuterVolumeSpecName: "utilities") pod "90267e5e-ba3a-46ed-90b7-aba5505cc38a" (UID: "90267e5e-ba3a-46ed-90b7-aba5505cc38a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:51:20 crc kubenswrapper[4873]: I1201 09:51:20.494235 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90267e5e-ba3a-46ed-90b7-aba5505cc38a-kube-api-access-phk2p" (OuterVolumeSpecName: "kube-api-access-phk2p") pod "90267e5e-ba3a-46ed-90b7-aba5505cc38a" (UID: "90267e5e-ba3a-46ed-90b7-aba5505cc38a"). InnerVolumeSpecName "kube-api-access-phk2p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:51:20 crc kubenswrapper[4873]: I1201 09:51:20.541125 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90267e5e-ba3a-46ed-90b7-aba5505cc38a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "90267e5e-ba3a-46ed-90b7-aba5505cc38a" (UID: "90267e5e-ba3a-46ed-90b7-aba5505cc38a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:51:20 crc kubenswrapper[4873]: I1201 09:51:20.588656 4873 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90267e5e-ba3a-46ed-90b7-aba5505cc38a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 09:51:20 crc kubenswrapper[4873]: I1201 09:51:20.588698 4873 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90267e5e-ba3a-46ed-90b7-aba5505cc38a-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 09:51:20 crc kubenswrapper[4873]: I1201 09:51:20.588708 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-phk2p\" (UniqueName: \"kubernetes.io/projected/90267e5e-ba3a-46ed-90b7-aba5505cc38a-kube-api-access-phk2p\") on node \"crc\" DevicePath \"\"" Dec 01 09:51:20 crc kubenswrapper[4873]: I1201 09:51:20.679083 4873 generic.go:334] "Generic (PLEG): container finished" podID="90267e5e-ba3a-46ed-90b7-aba5505cc38a" containerID="c1802b08f8bafded322f30e650309fe3f61b90a8ead290d0230b966c0dd24ae5" exitCode=0 Dec 01 09:51:20 crc kubenswrapper[4873]: I1201 09:51:20.679130 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kj4zj" event={"ID":"90267e5e-ba3a-46ed-90b7-aba5505cc38a","Type":"ContainerDied","Data":"c1802b08f8bafded322f30e650309fe3f61b90a8ead290d0230b966c0dd24ae5"} Dec 01 09:51:20 crc kubenswrapper[4873]: I1201 09:51:20.679191 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kj4zj" Dec 01 09:51:20 crc kubenswrapper[4873]: I1201 09:51:20.679210 4873 scope.go:117] "RemoveContainer" containerID="c1802b08f8bafded322f30e650309fe3f61b90a8ead290d0230b966c0dd24ae5" Dec 01 09:51:20 crc kubenswrapper[4873]: I1201 09:51:20.679194 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kj4zj" event={"ID":"90267e5e-ba3a-46ed-90b7-aba5505cc38a","Type":"ContainerDied","Data":"7799074176e8722a28f9bc4f1273208a3d789b5683eded32f76655952623fddf"} Dec 01 09:51:20 crc kubenswrapper[4873]: I1201 09:51:20.720940 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kj4zj"] Dec 01 09:51:20 crc kubenswrapper[4873]: I1201 09:51:20.722059 4873 scope.go:117] "RemoveContainer" containerID="6eb2b36c8f3d61d24ac4b0ca7e1301532f406424c8323865fec4f4f245e7227b" Dec 01 09:51:20 crc kubenswrapper[4873]: I1201 09:51:20.731515 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kj4zj"] Dec 01 09:51:20 crc kubenswrapper[4873]: I1201 09:51:20.745913 4873 scope.go:117] "RemoveContainer" containerID="88edda7a12b2b2e25cbae08427faf9603f918d08d722e7f3f097bdcb12bc53ad" Dec 01 09:51:20 crc kubenswrapper[4873]: I1201 09:51:20.795122 4873 scope.go:117] "RemoveContainer" containerID="c1802b08f8bafded322f30e650309fe3f61b90a8ead290d0230b966c0dd24ae5" Dec 01 09:51:20 crc kubenswrapper[4873]: E1201 09:51:20.796068 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1802b08f8bafded322f30e650309fe3f61b90a8ead290d0230b966c0dd24ae5\": container with ID starting with c1802b08f8bafded322f30e650309fe3f61b90a8ead290d0230b966c0dd24ae5 not found: ID does not exist" containerID="c1802b08f8bafded322f30e650309fe3f61b90a8ead290d0230b966c0dd24ae5" Dec 01 09:51:20 crc kubenswrapper[4873]: I1201 09:51:20.796137 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1802b08f8bafded322f30e650309fe3f61b90a8ead290d0230b966c0dd24ae5"} err="failed to get container status \"c1802b08f8bafded322f30e650309fe3f61b90a8ead290d0230b966c0dd24ae5\": rpc error: code = NotFound desc = could not find container \"c1802b08f8bafded322f30e650309fe3f61b90a8ead290d0230b966c0dd24ae5\": container with ID starting with c1802b08f8bafded322f30e650309fe3f61b90a8ead290d0230b966c0dd24ae5 not found: ID does not exist" Dec 01 09:51:20 crc kubenswrapper[4873]: I1201 09:51:20.796162 4873 scope.go:117] "RemoveContainer" containerID="6eb2b36c8f3d61d24ac4b0ca7e1301532f406424c8323865fec4f4f245e7227b" Dec 01 09:51:20 crc kubenswrapper[4873]: E1201 09:51:20.796626 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6eb2b36c8f3d61d24ac4b0ca7e1301532f406424c8323865fec4f4f245e7227b\": container with ID starting with 6eb2b36c8f3d61d24ac4b0ca7e1301532f406424c8323865fec4f4f245e7227b not found: ID does not exist" containerID="6eb2b36c8f3d61d24ac4b0ca7e1301532f406424c8323865fec4f4f245e7227b" Dec 01 09:51:20 crc kubenswrapper[4873]: I1201 09:51:20.796658 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6eb2b36c8f3d61d24ac4b0ca7e1301532f406424c8323865fec4f4f245e7227b"} err="failed to get container status \"6eb2b36c8f3d61d24ac4b0ca7e1301532f406424c8323865fec4f4f245e7227b\": rpc error: code = NotFound desc = could not find container \"6eb2b36c8f3d61d24ac4b0ca7e1301532f406424c8323865fec4f4f245e7227b\": container with ID starting with 6eb2b36c8f3d61d24ac4b0ca7e1301532f406424c8323865fec4f4f245e7227b not found: ID does not exist" Dec 01 09:51:20 crc kubenswrapper[4873]: I1201 09:51:20.796676 4873 scope.go:117] "RemoveContainer" containerID="88edda7a12b2b2e25cbae08427faf9603f918d08d722e7f3f097bdcb12bc53ad" Dec 01 09:51:20 crc kubenswrapper[4873]: E1201 09:51:20.796927 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88edda7a12b2b2e25cbae08427faf9603f918d08d722e7f3f097bdcb12bc53ad\": container with ID starting with 88edda7a12b2b2e25cbae08427faf9603f918d08d722e7f3f097bdcb12bc53ad not found: ID does not exist" containerID="88edda7a12b2b2e25cbae08427faf9603f918d08d722e7f3f097bdcb12bc53ad" Dec 01 09:51:20 crc kubenswrapper[4873]: I1201 09:51:20.796950 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88edda7a12b2b2e25cbae08427faf9603f918d08d722e7f3f097bdcb12bc53ad"} err="failed to get container status \"88edda7a12b2b2e25cbae08427faf9603f918d08d722e7f3f097bdcb12bc53ad\": rpc error: code = NotFound desc = could not find container \"88edda7a12b2b2e25cbae08427faf9603f918d08d722e7f3f097bdcb12bc53ad\": container with ID starting with 88edda7a12b2b2e25cbae08427faf9603f918d08d722e7f3f097bdcb12bc53ad not found: ID does not exist" Dec 01 09:51:22 crc kubenswrapper[4873]: I1201 09:51:22.450356 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90267e5e-ba3a-46ed-90b7-aba5505cc38a" path="/var/lib/kubelet/pods/90267e5e-ba3a-46ed-90b7-aba5505cc38a/volumes" Dec 01 09:51:31 crc kubenswrapper[4873]: I1201 09:51:31.058770 4873 patch_prober.go:28] interesting pod/machine-config-daemon-scwpp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:51:31 crc kubenswrapper[4873]: I1201 09:51:31.059485 4873 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:51:31 crc kubenswrapper[4873]: I1201 09:51:31.059554 4873 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" Dec 01 09:51:31 crc kubenswrapper[4873]: I1201 09:51:31.060600 4873 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"67abbbfa1eead10a18ea3d3104e50a98c59b5983a7e8d4be64c2d03f882521a7"} pod="openshift-machine-config-operator/machine-config-daemon-scwpp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 09:51:31 crc kubenswrapper[4873]: I1201 09:51:31.060777 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" containerName="machine-config-daemon" containerID="cri-o://67abbbfa1eead10a18ea3d3104e50a98c59b5983a7e8d4be64c2d03f882521a7" gracePeriod=600 Dec 01 09:51:31 crc kubenswrapper[4873]: E1201 09:51:31.192802 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:51:31 crc kubenswrapper[4873]: I1201 09:51:31.793607 4873 generic.go:334] "Generic (PLEG): container finished" podID="fef7b114-0e07-402d-a37b-315c36011f4b" containerID="67abbbfa1eead10a18ea3d3104e50a98c59b5983a7e8d4be64c2d03f882521a7" exitCode=0 Dec 01 09:51:31 crc kubenswrapper[4873]: I1201 09:51:31.793675 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" event={"ID":"fef7b114-0e07-402d-a37b-315c36011f4b","Type":"ContainerDied","Data":"67abbbfa1eead10a18ea3d3104e50a98c59b5983a7e8d4be64c2d03f882521a7"} Dec 01 09:51:31 crc kubenswrapper[4873]: I1201 09:51:31.793755 4873 scope.go:117] "RemoveContainer" containerID="29490d89d4ce9ae938873b01532387c61bd2469f197fb97051e2685e7e6f8a4e" Dec 01 09:51:31 crc kubenswrapper[4873]: I1201 09:51:31.794250 4873 scope.go:117] "RemoveContainer" containerID="67abbbfa1eead10a18ea3d3104e50a98c59b5983a7e8d4be64c2d03f882521a7" Dec 01 09:51:31 crc kubenswrapper[4873]: E1201 09:51:31.794656 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:51:34 crc kubenswrapper[4873]: I1201 09:51:34.012146 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhgjjb_b77a8d6f-5a83-494e-a366-300ffde2afb4/util/0.log" Dec 01 09:51:34 crc kubenswrapper[4873]: I1201 09:51:34.180405 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhgjjb_b77a8d6f-5a83-494e-a366-300ffde2afb4/pull/0.log" Dec 01 09:51:34 crc kubenswrapper[4873]: I1201 09:51:34.189225 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhgjjb_b77a8d6f-5a83-494e-a366-300ffde2afb4/util/0.log" Dec 01 09:51:34 crc kubenswrapper[4873]: I1201 09:51:34.249534 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhgjjb_b77a8d6f-5a83-494e-a366-300ffde2afb4/pull/0.log" Dec 01 09:51:34 crc kubenswrapper[4873]: I1201 09:51:34.438048 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhgjjb_b77a8d6f-5a83-494e-a366-300ffde2afb4/pull/0.log" Dec 01 09:51:34 crc kubenswrapper[4873]: I1201 09:51:34.442342 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhgjjb_b77a8d6f-5a83-494e-a366-300ffde2afb4/util/0.log" Dec 01 09:51:34 crc kubenswrapper[4873]: I1201 09:51:34.465509 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhgjjb_b77a8d6f-5a83-494e-a366-300ffde2afb4/extract/0.log" Dec 01 09:51:34 crc kubenswrapper[4873]: I1201 09:51:34.620250 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8359xjq_8d9127f2-7870-4647-8655-0c8fff4df500/util/0.log" Dec 01 09:51:34 crc kubenswrapper[4873]: I1201 09:51:34.852589 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8359xjq_8d9127f2-7870-4647-8655-0c8fff4df500/util/0.log" Dec 01 09:51:34 crc kubenswrapper[4873]: I1201 09:51:34.863400 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8359xjq_8d9127f2-7870-4647-8655-0c8fff4df500/pull/0.log" Dec 01 09:51:34 crc kubenswrapper[4873]: I1201 09:51:34.894241 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8359xjq_8d9127f2-7870-4647-8655-0c8fff4df500/pull/0.log" Dec 01 09:51:35 crc kubenswrapper[4873]: I1201 09:51:35.052186 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8359xjq_8d9127f2-7870-4647-8655-0c8fff4df500/pull/0.log" Dec 01 09:51:35 crc kubenswrapper[4873]: I1201 09:51:35.095768 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8359xjq_8d9127f2-7870-4647-8655-0c8fff4df500/extract/0.log" Dec 01 09:51:35 crc kubenswrapper[4873]: I1201 09:51:35.125252 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8359xjq_8d9127f2-7870-4647-8655-0c8fff4df500/util/0.log" Dec 01 09:51:35 crc kubenswrapper[4873]: I1201 09:51:35.272926 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pdgt5_96e1f659-b44b-4573-ba28-9605403487e5/extract-utilities/0.log" Dec 01 09:51:35 crc kubenswrapper[4873]: I1201 09:51:35.514031 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pdgt5_96e1f659-b44b-4573-ba28-9605403487e5/extract-utilities/0.log" Dec 01 09:51:35 crc kubenswrapper[4873]: I1201 09:51:35.522952 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pdgt5_96e1f659-b44b-4573-ba28-9605403487e5/extract-content/0.log" Dec 01 09:51:35 crc kubenswrapper[4873]: I1201 09:51:35.527359 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pdgt5_96e1f659-b44b-4573-ba28-9605403487e5/extract-content/0.log" Dec 01 09:51:35 crc kubenswrapper[4873]: I1201 09:51:35.698116 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pdgt5_96e1f659-b44b-4573-ba28-9605403487e5/extract-utilities/0.log" Dec 01 09:51:35 crc kubenswrapper[4873]: I1201 09:51:35.812087 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pdgt5_96e1f659-b44b-4573-ba28-9605403487e5/extract-content/0.log" Dec 01 09:51:36 crc kubenswrapper[4873]: I1201 09:51:36.200507 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7zq5j_3adbfc1d-7e40-4bfc-ab02-bcec5b0a5c38/extract-utilities/0.log" Dec 01 09:51:36 crc kubenswrapper[4873]: I1201 09:51:36.494326 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7zq5j_3adbfc1d-7e40-4bfc-ab02-bcec5b0a5c38/extract-utilities/0.log" Dec 01 09:51:36 crc kubenswrapper[4873]: I1201 09:51:36.553862 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7zq5j_3adbfc1d-7e40-4bfc-ab02-bcec5b0a5c38/extract-content/0.log" Dec 01 09:51:36 crc kubenswrapper[4873]: I1201 09:51:36.611401 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pdgt5_96e1f659-b44b-4573-ba28-9605403487e5/registry-server/0.log" Dec 01 09:51:36 crc kubenswrapper[4873]: I1201 09:51:36.614472 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7zq5j_3adbfc1d-7e40-4bfc-ab02-bcec5b0a5c38/extract-content/0.log" Dec 01 09:51:36 crc kubenswrapper[4873]: I1201 09:51:36.760174 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7zq5j_3adbfc1d-7e40-4bfc-ab02-bcec5b0a5c38/extract-utilities/0.log" Dec 01 09:51:36 crc kubenswrapper[4873]: I1201 09:51:36.853652 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7zq5j_3adbfc1d-7e40-4bfc-ab02-bcec5b0a5c38/extract-content/0.log" Dec 01 09:51:37 crc kubenswrapper[4873]: I1201 09:51:37.130187 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-dgdcc_c0e7b6b5-8852-4dec-bbf2-b7247b34a791/marketplace-operator/0.log" Dec 01 09:51:37 crc kubenswrapper[4873]: I1201 09:51:37.166628 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7zq5j_3adbfc1d-7e40-4bfc-ab02-bcec5b0a5c38/registry-server/0.log" Dec 01 09:51:37 crc kubenswrapper[4873]: I1201 09:51:37.292137 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-p98sp_d62ca400-6749-4a3f-94ec-f7a5716584b9/extract-utilities/0.log" Dec 01 09:51:37 crc kubenswrapper[4873]: I1201 09:51:37.466729 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-p98sp_d62ca400-6749-4a3f-94ec-f7a5716584b9/extract-utilities/0.log" Dec 01 09:51:37 crc kubenswrapper[4873]: I1201 09:51:37.512123 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-p98sp_d62ca400-6749-4a3f-94ec-f7a5716584b9/extract-content/0.log" Dec 01 09:51:37 crc kubenswrapper[4873]: I1201 09:51:37.525363 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-p98sp_d62ca400-6749-4a3f-94ec-f7a5716584b9/extract-content/0.log" Dec 01 09:51:37 crc kubenswrapper[4873]: I1201 09:51:37.763792 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-p98sp_d62ca400-6749-4a3f-94ec-f7a5716584b9/extract-content/0.log" Dec 01 09:51:37 crc kubenswrapper[4873]: I1201 09:51:37.765622 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-p98sp_d62ca400-6749-4a3f-94ec-f7a5716584b9/extract-utilities/0.log" Dec 01 09:51:37 crc kubenswrapper[4873]: I1201 09:51:37.955838 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-p98sp_d62ca400-6749-4a3f-94ec-f7a5716584b9/registry-server/0.log" Dec 01 09:51:38 crc kubenswrapper[4873]: I1201 09:51:38.020523 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-v6fx2_3e92483b-2415-4432-9059-3a60badb2e4a/extract-utilities/0.log" Dec 01 09:51:38 crc kubenswrapper[4873]: I1201 09:51:38.205236 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-v6fx2_3e92483b-2415-4432-9059-3a60badb2e4a/extract-utilities/0.log" Dec 01 09:51:38 crc kubenswrapper[4873]: I1201 09:51:38.243544 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-v6fx2_3e92483b-2415-4432-9059-3a60badb2e4a/extract-content/0.log" Dec 01 09:51:38 crc kubenswrapper[4873]: I1201 09:51:38.257716 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-v6fx2_3e92483b-2415-4432-9059-3a60badb2e4a/extract-content/0.log" Dec 01 09:51:38 crc kubenswrapper[4873]: I1201 09:51:38.495760 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-v6fx2_3e92483b-2415-4432-9059-3a60badb2e4a/extract-content/0.log" Dec 01 09:51:38 crc kubenswrapper[4873]: I1201 09:51:38.519349 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-v6fx2_3e92483b-2415-4432-9059-3a60badb2e4a/extract-utilities/0.log" Dec 01 09:51:39 crc kubenswrapper[4873]: I1201 09:51:39.171739 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-v6fx2_3e92483b-2415-4432-9059-3a60badb2e4a/registry-server/0.log" Dec 01 09:51:46 crc kubenswrapper[4873]: I1201 09:51:46.430178 4873 scope.go:117] "RemoveContainer" containerID="67abbbfa1eead10a18ea3d3104e50a98c59b5983a7e8d4be64c2d03f882521a7" Dec 01 09:51:46 crc kubenswrapper[4873]: E1201 09:51:46.431109 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:51:51 crc kubenswrapper[4873]: I1201 09:51:51.779943 4873 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-825sq"] Dec 01 09:51:51 crc kubenswrapper[4873]: E1201 09:51:51.781283 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90267e5e-ba3a-46ed-90b7-aba5505cc38a" containerName="extract-content" Dec 01 09:51:51 crc kubenswrapper[4873]: I1201 09:51:51.781304 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="90267e5e-ba3a-46ed-90b7-aba5505cc38a" containerName="extract-content" Dec 01 09:51:51 crc kubenswrapper[4873]: E1201 09:51:51.781401 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90267e5e-ba3a-46ed-90b7-aba5505cc38a" containerName="extract-utilities" Dec 01 09:51:51 crc kubenswrapper[4873]: I1201 09:51:51.781410 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="90267e5e-ba3a-46ed-90b7-aba5505cc38a" containerName="extract-utilities" Dec 01 09:51:51 crc kubenswrapper[4873]: E1201 09:51:51.781425 4873 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90267e5e-ba3a-46ed-90b7-aba5505cc38a" containerName="registry-server" Dec 01 09:51:51 crc kubenswrapper[4873]: I1201 09:51:51.781437 4873 state_mem.go:107] "Deleted CPUSet assignment" podUID="90267e5e-ba3a-46ed-90b7-aba5505cc38a" containerName="registry-server" Dec 01 09:51:51 crc kubenswrapper[4873]: I1201 09:51:51.781691 4873 memory_manager.go:354] "RemoveStaleState removing state" podUID="90267e5e-ba3a-46ed-90b7-aba5505cc38a" containerName="registry-server" Dec 01 09:51:51 crc kubenswrapper[4873]: I1201 09:51:51.790384 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-825sq" Dec 01 09:51:51 crc kubenswrapper[4873]: I1201 09:51:51.806425 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-825sq"] Dec 01 09:51:51 crc kubenswrapper[4873]: I1201 09:51:51.906856 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb-catalog-content\") pod \"redhat-operators-825sq\" (UID: \"7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb\") " pod="openshift-marketplace/redhat-operators-825sq" Dec 01 09:51:51 crc kubenswrapper[4873]: I1201 09:51:51.907364 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb-utilities\") pod \"redhat-operators-825sq\" (UID: \"7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb\") " pod="openshift-marketplace/redhat-operators-825sq" Dec 01 09:51:51 crc kubenswrapper[4873]: I1201 09:51:51.907475 4873 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hvvg\" (UniqueName: \"kubernetes.io/projected/7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb-kube-api-access-6hvvg\") pod \"redhat-operators-825sq\" (UID: \"7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb\") " pod="openshift-marketplace/redhat-operators-825sq" Dec 01 09:51:52 crc kubenswrapper[4873]: I1201 09:51:52.010211 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb-catalog-content\") pod \"redhat-operators-825sq\" (UID: \"7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb\") " pod="openshift-marketplace/redhat-operators-825sq" Dec 01 09:51:52 crc kubenswrapper[4873]: I1201 09:51:52.011030 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb-catalog-content\") pod \"redhat-operators-825sq\" (UID: \"7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb\") " pod="openshift-marketplace/redhat-operators-825sq" Dec 01 09:51:52 crc kubenswrapper[4873]: I1201 09:51:52.011684 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb-utilities\") pod \"redhat-operators-825sq\" (UID: \"7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb\") " pod="openshift-marketplace/redhat-operators-825sq" Dec 01 09:51:52 crc kubenswrapper[4873]: I1201 09:51:52.012185 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb-utilities\") pod \"redhat-operators-825sq\" (UID: \"7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb\") " pod="openshift-marketplace/redhat-operators-825sq" Dec 01 09:51:52 crc kubenswrapper[4873]: I1201 09:51:52.011718 4873 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hvvg\" (UniqueName: \"kubernetes.io/projected/7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb-kube-api-access-6hvvg\") pod \"redhat-operators-825sq\" (UID: \"7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb\") " pod="openshift-marketplace/redhat-operators-825sq" Dec 01 09:51:52 crc kubenswrapper[4873]: I1201 09:51:52.048432 4873 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hvvg\" (UniqueName: \"kubernetes.io/projected/7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb-kube-api-access-6hvvg\") pod \"redhat-operators-825sq\" (UID: \"7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb\") " pod="openshift-marketplace/redhat-operators-825sq" Dec 01 09:51:52 crc kubenswrapper[4873]: I1201 09:51:52.126398 4873 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-825sq" Dec 01 09:51:52 crc kubenswrapper[4873]: I1201 09:51:52.690942 4873 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-825sq"] Dec 01 09:51:53 crc kubenswrapper[4873]: I1201 09:51:53.051206 4873 generic.go:334] "Generic (PLEG): container finished" podID="7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb" containerID="bbf56cb44f1409045a144618167248a1bb674af445ad5e4fa57dffd5f6c54d1c" exitCode=0 Dec 01 09:51:53 crc kubenswrapper[4873]: I1201 09:51:53.051781 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-825sq" event={"ID":"7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb","Type":"ContainerDied","Data":"bbf56cb44f1409045a144618167248a1bb674af445ad5e4fa57dffd5f6c54d1c"} Dec 01 09:51:53 crc kubenswrapper[4873]: I1201 09:51:53.051847 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-825sq" event={"ID":"7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb","Type":"ContainerStarted","Data":"4fb203d5a44ddd54e33f29e330ce263c14636398a76fab92bce2ccee77806d4a"} Dec 01 09:51:54 crc kubenswrapper[4873]: I1201 09:51:54.063047 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-825sq" event={"ID":"7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb","Type":"ContainerStarted","Data":"d5b12de6001661ddf4900c7ecbd0e09f9aea4d72415b021c0eb9a5dcb9987d02"} Dec 01 09:51:55 crc kubenswrapper[4873]: I1201 09:51:55.075255 4873 generic.go:334] "Generic (PLEG): container finished" podID="7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb" containerID="d5b12de6001661ddf4900c7ecbd0e09f9aea4d72415b021c0eb9a5dcb9987d02" exitCode=0 Dec 01 09:51:55 crc kubenswrapper[4873]: I1201 09:51:55.075442 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-825sq" event={"ID":"7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb","Type":"ContainerDied","Data":"d5b12de6001661ddf4900c7ecbd0e09f9aea4d72415b021c0eb9a5dcb9987d02"} Dec 01 09:51:56 crc kubenswrapper[4873]: I1201 09:51:56.088540 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-825sq" event={"ID":"7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb","Type":"ContainerStarted","Data":"6189754fcd0adfca1085fd4cf1a6c1585216df20a9622a3a530c0a2587e1f561"} Dec 01 09:51:56 crc kubenswrapper[4873]: I1201 09:51:56.110984 4873 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-825sq" podStartSLOduration=2.341077042 podStartE2EDuration="5.110947892s" podCreationTimestamp="2025-12-01 09:51:51 +0000 UTC" firstStartedPulling="2025-12-01 09:51:53.053701732 +0000 UTC m=+4288.955810271" lastFinishedPulling="2025-12-01 09:51:55.823572582 +0000 UTC m=+4291.725681121" observedRunningTime="2025-12-01 09:51:56.108274624 +0000 UTC m=+4292.010383183" watchObservedRunningTime="2025-12-01 09:51:56.110947892 +0000 UTC m=+4292.013056431" Dec 01 09:51:57 crc kubenswrapper[4873]: I1201 09:51:57.430527 4873 scope.go:117] "RemoveContainer" containerID="67abbbfa1eead10a18ea3d3104e50a98c59b5983a7e8d4be64c2d03f882521a7" Dec 01 09:51:57 crc kubenswrapper[4873]: E1201 09:51:57.431203 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:52:02 crc kubenswrapper[4873]: I1201 09:52:02.127271 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-825sq" Dec 01 09:52:02 crc kubenswrapper[4873]: I1201 09:52:02.128067 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-825sq" Dec 01 09:52:02 crc kubenswrapper[4873]: I1201 09:52:02.203613 4873 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-825sq" Dec 01 09:52:02 crc kubenswrapper[4873]: I1201 09:52:02.267477 4873 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-825sq" Dec 01 09:52:02 crc kubenswrapper[4873]: I1201 09:52:02.450095 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-825sq"] Dec 01 09:52:04 crc kubenswrapper[4873]: I1201 09:52:04.198860 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-825sq" podUID="7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb" containerName="registry-server" containerID="cri-o://6189754fcd0adfca1085fd4cf1a6c1585216df20a9622a3a530c0a2587e1f561" gracePeriod=2 Dec 01 09:52:04 crc kubenswrapper[4873]: I1201 09:52:04.776590 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-825sq" Dec 01 09:52:04 crc kubenswrapper[4873]: I1201 09:52:04.909460 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb-utilities\") pod \"7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb\" (UID: \"7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb\") " Dec 01 09:52:04 crc kubenswrapper[4873]: I1201 09:52:04.909574 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb-catalog-content\") pod \"7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb\" (UID: \"7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb\") " Dec 01 09:52:04 crc kubenswrapper[4873]: I1201 09:52:04.909712 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6hvvg\" (UniqueName: \"kubernetes.io/projected/7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb-kube-api-access-6hvvg\") pod \"7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb\" (UID: \"7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb\") " Dec 01 09:52:04 crc kubenswrapper[4873]: I1201 09:52:04.910332 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb-utilities" (OuterVolumeSpecName: "utilities") pod "7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb" (UID: "7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:52:04 crc kubenswrapper[4873]: I1201 09:52:04.911492 4873 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 09:52:04 crc kubenswrapper[4873]: I1201 09:52:04.918226 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb-kube-api-access-6hvvg" (OuterVolumeSpecName: "kube-api-access-6hvvg") pod "7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb" (UID: "7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb"). InnerVolumeSpecName "kube-api-access-6hvvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:52:05 crc kubenswrapper[4873]: I1201 09:52:05.014480 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6hvvg\" (UniqueName: \"kubernetes.io/projected/7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb-kube-api-access-6hvvg\") on node \"crc\" DevicePath \"\"" Dec 01 09:52:05 crc kubenswrapper[4873]: I1201 09:52:05.038600 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb" (UID: "7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:52:05 crc kubenswrapper[4873]: I1201 09:52:05.116377 4873 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 09:52:05 crc kubenswrapper[4873]: I1201 09:52:05.210249 4873 generic.go:334] "Generic (PLEG): container finished" podID="7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb" containerID="6189754fcd0adfca1085fd4cf1a6c1585216df20a9622a3a530c0a2587e1f561" exitCode=0 Dec 01 09:52:05 crc kubenswrapper[4873]: I1201 09:52:05.210306 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-825sq" event={"ID":"7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb","Type":"ContainerDied","Data":"6189754fcd0adfca1085fd4cf1a6c1585216df20a9622a3a530c0a2587e1f561"} Dec 01 09:52:05 crc kubenswrapper[4873]: I1201 09:52:05.210347 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-825sq" event={"ID":"7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb","Type":"ContainerDied","Data":"4fb203d5a44ddd54e33f29e330ce263c14636398a76fab92bce2ccee77806d4a"} Dec 01 09:52:05 crc kubenswrapper[4873]: I1201 09:52:05.210371 4873 scope.go:117] "RemoveContainer" containerID="6189754fcd0adfca1085fd4cf1a6c1585216df20a9622a3a530c0a2587e1f561" Dec 01 09:52:05 crc kubenswrapper[4873]: I1201 09:52:05.210400 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-825sq" Dec 01 09:52:05 crc kubenswrapper[4873]: I1201 09:52:05.252550 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-825sq"] Dec 01 09:52:05 crc kubenswrapper[4873]: I1201 09:52:05.257237 4873 scope.go:117] "RemoveContainer" containerID="d5b12de6001661ddf4900c7ecbd0e09f9aea4d72415b021c0eb9a5dcb9987d02" Dec 01 09:52:05 crc kubenswrapper[4873]: I1201 09:52:05.266589 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-825sq"] Dec 01 09:52:05 crc kubenswrapper[4873]: I1201 09:52:05.290587 4873 scope.go:117] "RemoveContainer" containerID="bbf56cb44f1409045a144618167248a1bb674af445ad5e4fa57dffd5f6c54d1c" Dec 01 09:52:05 crc kubenswrapper[4873]: I1201 09:52:05.334802 4873 scope.go:117] "RemoveContainer" containerID="6189754fcd0adfca1085fd4cf1a6c1585216df20a9622a3a530c0a2587e1f561" Dec 01 09:52:05 crc kubenswrapper[4873]: E1201 09:52:05.335394 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6189754fcd0adfca1085fd4cf1a6c1585216df20a9622a3a530c0a2587e1f561\": container with ID starting with 6189754fcd0adfca1085fd4cf1a6c1585216df20a9622a3a530c0a2587e1f561 not found: ID does not exist" containerID="6189754fcd0adfca1085fd4cf1a6c1585216df20a9622a3a530c0a2587e1f561" Dec 01 09:52:05 crc kubenswrapper[4873]: I1201 09:52:05.335463 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6189754fcd0adfca1085fd4cf1a6c1585216df20a9622a3a530c0a2587e1f561"} err="failed to get container status \"6189754fcd0adfca1085fd4cf1a6c1585216df20a9622a3a530c0a2587e1f561\": rpc error: code = NotFound desc = could not find container \"6189754fcd0adfca1085fd4cf1a6c1585216df20a9622a3a530c0a2587e1f561\": container with ID starting with 6189754fcd0adfca1085fd4cf1a6c1585216df20a9622a3a530c0a2587e1f561 not found: ID does not exist" Dec 01 09:52:05 crc kubenswrapper[4873]: I1201 09:52:05.335504 4873 scope.go:117] "RemoveContainer" containerID="d5b12de6001661ddf4900c7ecbd0e09f9aea4d72415b021c0eb9a5dcb9987d02" Dec 01 09:52:05 crc kubenswrapper[4873]: E1201 09:52:05.338398 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5b12de6001661ddf4900c7ecbd0e09f9aea4d72415b021c0eb9a5dcb9987d02\": container with ID starting with d5b12de6001661ddf4900c7ecbd0e09f9aea4d72415b021c0eb9a5dcb9987d02 not found: ID does not exist" containerID="d5b12de6001661ddf4900c7ecbd0e09f9aea4d72415b021c0eb9a5dcb9987d02" Dec 01 09:52:05 crc kubenswrapper[4873]: I1201 09:52:05.338455 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5b12de6001661ddf4900c7ecbd0e09f9aea4d72415b021c0eb9a5dcb9987d02"} err="failed to get container status \"d5b12de6001661ddf4900c7ecbd0e09f9aea4d72415b021c0eb9a5dcb9987d02\": rpc error: code = NotFound desc = could not find container \"d5b12de6001661ddf4900c7ecbd0e09f9aea4d72415b021c0eb9a5dcb9987d02\": container with ID starting with d5b12de6001661ddf4900c7ecbd0e09f9aea4d72415b021c0eb9a5dcb9987d02 not found: ID does not exist" Dec 01 09:52:05 crc kubenswrapper[4873]: I1201 09:52:05.338495 4873 scope.go:117] "RemoveContainer" containerID="bbf56cb44f1409045a144618167248a1bb674af445ad5e4fa57dffd5f6c54d1c" Dec 01 09:52:05 crc kubenswrapper[4873]: E1201 09:52:05.339094 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bbf56cb44f1409045a144618167248a1bb674af445ad5e4fa57dffd5f6c54d1c\": container with ID starting with bbf56cb44f1409045a144618167248a1bb674af445ad5e4fa57dffd5f6c54d1c not found: ID does not exist" containerID="bbf56cb44f1409045a144618167248a1bb674af445ad5e4fa57dffd5f6c54d1c" Dec 01 09:52:05 crc kubenswrapper[4873]: I1201 09:52:05.339133 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbf56cb44f1409045a144618167248a1bb674af445ad5e4fa57dffd5f6c54d1c"} err="failed to get container status \"bbf56cb44f1409045a144618167248a1bb674af445ad5e4fa57dffd5f6c54d1c\": rpc error: code = NotFound desc = could not find container \"bbf56cb44f1409045a144618167248a1bb674af445ad5e4fa57dffd5f6c54d1c\": container with ID starting with bbf56cb44f1409045a144618167248a1bb674af445ad5e4fa57dffd5f6c54d1c not found: ID does not exist" Dec 01 09:52:06 crc kubenswrapper[4873]: I1201 09:52:06.447688 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb" path="/var/lib/kubelet/pods/7c3f6db9-f4dd-40a0-8a7c-b23503bd03cb/volumes" Dec 01 09:52:10 crc kubenswrapper[4873]: I1201 09:52:10.431128 4873 scope.go:117] "RemoveContainer" containerID="67abbbfa1eead10a18ea3d3104e50a98c59b5983a7e8d4be64c2d03f882521a7" Dec 01 09:52:10 crc kubenswrapper[4873]: E1201 09:52:10.431871 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:52:24 crc kubenswrapper[4873]: I1201 09:52:24.438615 4873 scope.go:117] "RemoveContainer" containerID="67abbbfa1eead10a18ea3d3104e50a98c59b5983a7e8d4be64c2d03f882521a7" Dec 01 09:52:24 crc kubenswrapper[4873]: E1201 09:52:24.440180 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:52:38 crc kubenswrapper[4873]: I1201 09:52:38.430626 4873 scope.go:117] "RemoveContainer" containerID="67abbbfa1eead10a18ea3d3104e50a98c59b5983a7e8d4be64c2d03f882521a7" Dec 01 09:52:38 crc kubenswrapper[4873]: E1201 09:52:38.431516 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:52:49 crc kubenswrapper[4873]: I1201 09:52:49.431614 4873 scope.go:117] "RemoveContainer" containerID="67abbbfa1eead10a18ea3d3104e50a98c59b5983a7e8d4be64c2d03f882521a7" Dec 01 09:52:49 crc kubenswrapper[4873]: E1201 09:52:49.432554 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:53:01 crc kubenswrapper[4873]: I1201 09:53:01.432713 4873 scope.go:117] "RemoveContainer" containerID="67abbbfa1eead10a18ea3d3104e50a98c59b5983a7e8d4be64c2d03f882521a7" Dec 01 09:53:01 crc kubenswrapper[4873]: E1201 09:53:01.433788 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:53:13 crc kubenswrapper[4873]: I1201 09:53:13.429846 4873 scope.go:117] "RemoveContainer" containerID="67abbbfa1eead10a18ea3d3104e50a98c59b5983a7e8d4be64c2d03f882521a7" Dec 01 09:53:13 crc kubenswrapper[4873]: E1201 09:53:13.430696 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:53:26 crc kubenswrapper[4873]: I1201 09:53:26.431363 4873 scope.go:117] "RemoveContainer" containerID="67abbbfa1eead10a18ea3d3104e50a98c59b5983a7e8d4be64c2d03f882521a7" Dec 01 09:53:26 crc kubenswrapper[4873]: E1201 09:53:26.433635 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:53:38 crc kubenswrapper[4873]: I1201 09:53:38.437768 4873 scope.go:117] "RemoveContainer" containerID="67abbbfa1eead10a18ea3d3104e50a98c59b5983a7e8d4be64c2d03f882521a7" Dec 01 09:53:38 crc kubenswrapper[4873]: E1201 09:53:38.438739 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:53:42 crc kubenswrapper[4873]: I1201 09:53:42.204117 4873 generic.go:334] "Generic (PLEG): container finished" podID="e7f569d0-eb50-4458-9144-b8f4c8947bf1" containerID="c4c09f5d02e8d778987bf1b9d67ee56115a1328f9ae89c8890d6ac9e77e66ad6" exitCode=0 Dec 01 09:53:42 crc kubenswrapper[4873]: I1201 09:53:42.204221 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mxwtf/must-gather-gbtqv" event={"ID":"e7f569d0-eb50-4458-9144-b8f4c8947bf1","Type":"ContainerDied","Data":"c4c09f5d02e8d778987bf1b9d67ee56115a1328f9ae89c8890d6ac9e77e66ad6"} Dec 01 09:53:42 crc kubenswrapper[4873]: I1201 09:53:42.205373 4873 scope.go:117] "RemoveContainer" containerID="c4c09f5d02e8d778987bf1b9d67ee56115a1328f9ae89c8890d6ac9e77e66ad6" Dec 01 09:53:42 crc kubenswrapper[4873]: I1201 09:53:42.631966 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mxwtf_must-gather-gbtqv_e7f569d0-eb50-4458-9144-b8f4c8947bf1/gather/0.log" Dec 01 09:53:53 crc kubenswrapper[4873]: I1201 09:53:53.024494 4873 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mxwtf/must-gather-gbtqv"] Dec 01 09:53:53 crc kubenswrapper[4873]: I1201 09:53:53.025912 4873 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-mxwtf/must-gather-gbtqv" podUID="e7f569d0-eb50-4458-9144-b8f4c8947bf1" containerName="copy" containerID="cri-o://70ca5c7fe0c7a45d30368191d719364c4255c891f9037fa451ac33885a58fb1d" gracePeriod=2 Dec 01 09:53:53 crc kubenswrapper[4873]: I1201 09:53:53.045544 4873 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mxwtf/must-gather-gbtqv"] Dec 01 09:53:53 crc kubenswrapper[4873]: I1201 09:53:53.430606 4873 scope.go:117] "RemoveContainer" containerID="67abbbfa1eead10a18ea3d3104e50a98c59b5983a7e8d4be64c2d03f882521a7" Dec 01 09:53:53 crc kubenswrapper[4873]: E1201 09:53:53.431069 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:53:54 crc kubenswrapper[4873]: I1201 09:53:54.195837 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mxwtf_must-gather-gbtqv_e7f569d0-eb50-4458-9144-b8f4c8947bf1/copy/0.log" Dec 01 09:53:54 crc kubenswrapper[4873]: I1201 09:53:54.198063 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mxwtf/must-gather-gbtqv" Dec 01 09:53:54 crc kubenswrapper[4873]: I1201 09:53:54.266481 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrxw4\" (UniqueName: \"kubernetes.io/projected/e7f569d0-eb50-4458-9144-b8f4c8947bf1-kube-api-access-nrxw4\") pod \"e7f569d0-eb50-4458-9144-b8f4c8947bf1\" (UID: \"e7f569d0-eb50-4458-9144-b8f4c8947bf1\") " Dec 01 09:53:54 crc kubenswrapper[4873]: I1201 09:53:54.266554 4873 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e7f569d0-eb50-4458-9144-b8f4c8947bf1-must-gather-output\") pod \"e7f569d0-eb50-4458-9144-b8f4c8947bf1\" (UID: \"e7f569d0-eb50-4458-9144-b8f4c8947bf1\") " Dec 01 09:53:54 crc kubenswrapper[4873]: I1201 09:53:54.339438 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7f569d0-eb50-4458-9144-b8f4c8947bf1-kube-api-access-nrxw4" (OuterVolumeSpecName: "kube-api-access-nrxw4") pod "e7f569d0-eb50-4458-9144-b8f4c8947bf1" (UID: "e7f569d0-eb50-4458-9144-b8f4c8947bf1"). InnerVolumeSpecName "kube-api-access-nrxw4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:53:54 crc kubenswrapper[4873]: I1201 09:53:54.349602 4873 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mxwtf_must-gather-gbtqv_e7f569d0-eb50-4458-9144-b8f4c8947bf1/copy/0.log" Dec 01 09:53:54 crc kubenswrapper[4873]: I1201 09:53:54.349945 4873 generic.go:334] "Generic (PLEG): container finished" podID="e7f569d0-eb50-4458-9144-b8f4c8947bf1" containerID="70ca5c7fe0c7a45d30368191d719364c4255c891f9037fa451ac33885a58fb1d" exitCode=143 Dec 01 09:53:54 crc kubenswrapper[4873]: I1201 09:53:54.350006 4873 scope.go:117] "RemoveContainer" containerID="70ca5c7fe0c7a45d30368191d719364c4255c891f9037fa451ac33885a58fb1d" Dec 01 09:53:54 crc kubenswrapper[4873]: I1201 09:53:54.350226 4873 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mxwtf/must-gather-gbtqv" Dec 01 09:53:54 crc kubenswrapper[4873]: I1201 09:53:54.369979 4873 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrxw4\" (UniqueName: \"kubernetes.io/projected/e7f569d0-eb50-4458-9144-b8f4c8947bf1-kube-api-access-nrxw4\") on node \"crc\" DevicePath \"\"" Dec 01 09:53:54 crc kubenswrapper[4873]: I1201 09:53:54.379889 4873 scope.go:117] "RemoveContainer" containerID="c4c09f5d02e8d778987bf1b9d67ee56115a1328f9ae89c8890d6ac9e77e66ad6" Dec 01 09:53:54 crc kubenswrapper[4873]: I1201 09:53:54.499349 4873 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7f569d0-eb50-4458-9144-b8f4c8947bf1-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "e7f569d0-eb50-4458-9144-b8f4c8947bf1" (UID: "e7f569d0-eb50-4458-9144-b8f4c8947bf1"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:53:54 crc kubenswrapper[4873]: I1201 09:53:54.503651 4873 scope.go:117] "RemoveContainer" containerID="70ca5c7fe0c7a45d30368191d719364c4255c891f9037fa451ac33885a58fb1d" Dec 01 09:53:54 crc kubenswrapper[4873]: E1201 09:53:54.504496 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70ca5c7fe0c7a45d30368191d719364c4255c891f9037fa451ac33885a58fb1d\": container with ID starting with 70ca5c7fe0c7a45d30368191d719364c4255c891f9037fa451ac33885a58fb1d not found: ID does not exist" containerID="70ca5c7fe0c7a45d30368191d719364c4255c891f9037fa451ac33885a58fb1d" Dec 01 09:53:54 crc kubenswrapper[4873]: I1201 09:53:54.504543 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70ca5c7fe0c7a45d30368191d719364c4255c891f9037fa451ac33885a58fb1d"} err="failed to get container status \"70ca5c7fe0c7a45d30368191d719364c4255c891f9037fa451ac33885a58fb1d\": rpc error: code = NotFound desc = could not find container \"70ca5c7fe0c7a45d30368191d719364c4255c891f9037fa451ac33885a58fb1d\": container with ID starting with 70ca5c7fe0c7a45d30368191d719364c4255c891f9037fa451ac33885a58fb1d not found: ID does not exist" Dec 01 09:53:54 crc kubenswrapper[4873]: I1201 09:53:54.504602 4873 scope.go:117] "RemoveContainer" containerID="c4c09f5d02e8d778987bf1b9d67ee56115a1328f9ae89c8890d6ac9e77e66ad6" Dec 01 09:53:54 crc kubenswrapper[4873]: E1201 09:53:54.505379 4873 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4c09f5d02e8d778987bf1b9d67ee56115a1328f9ae89c8890d6ac9e77e66ad6\": container with ID starting with c4c09f5d02e8d778987bf1b9d67ee56115a1328f9ae89c8890d6ac9e77e66ad6 not found: ID does not exist" containerID="c4c09f5d02e8d778987bf1b9d67ee56115a1328f9ae89c8890d6ac9e77e66ad6" Dec 01 09:53:54 crc kubenswrapper[4873]: I1201 09:53:54.505411 4873 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4c09f5d02e8d778987bf1b9d67ee56115a1328f9ae89c8890d6ac9e77e66ad6"} err="failed to get container status \"c4c09f5d02e8d778987bf1b9d67ee56115a1328f9ae89c8890d6ac9e77e66ad6\": rpc error: code = NotFound desc = could not find container \"c4c09f5d02e8d778987bf1b9d67ee56115a1328f9ae89c8890d6ac9e77e66ad6\": container with ID starting with c4c09f5d02e8d778987bf1b9d67ee56115a1328f9ae89c8890d6ac9e77e66ad6 not found: ID does not exist" Dec 01 09:53:54 crc kubenswrapper[4873]: I1201 09:53:54.576364 4873 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e7f569d0-eb50-4458-9144-b8f4c8947bf1-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 01 09:53:56 crc kubenswrapper[4873]: I1201 09:53:56.451773 4873 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7f569d0-eb50-4458-9144-b8f4c8947bf1" path="/var/lib/kubelet/pods/e7f569d0-eb50-4458-9144-b8f4c8947bf1/volumes" Dec 01 09:54:04 crc kubenswrapper[4873]: I1201 09:54:04.437884 4873 scope.go:117] "RemoveContainer" containerID="67abbbfa1eead10a18ea3d3104e50a98c59b5983a7e8d4be64c2d03f882521a7" Dec 01 09:54:04 crc kubenswrapper[4873]: E1201 09:54:04.439602 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:54:06 crc kubenswrapper[4873]: I1201 09:54:06.536520 4873 scope.go:117] "RemoveContainer" containerID="0b69ff28cc69f58b8b0993eb581f455022cb677089eea31cc4197b0cc6dbad03" Dec 01 09:54:18 crc kubenswrapper[4873]: I1201 09:54:18.431173 4873 scope.go:117] "RemoveContainer" containerID="67abbbfa1eead10a18ea3d3104e50a98c59b5983a7e8d4be64c2d03f882521a7" Dec 01 09:54:18 crc kubenswrapper[4873]: E1201 09:54:18.432169 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:54:29 crc kubenswrapper[4873]: I1201 09:54:29.431712 4873 scope.go:117] "RemoveContainer" containerID="67abbbfa1eead10a18ea3d3104e50a98c59b5983a7e8d4be64c2d03f882521a7" Dec 01 09:54:29 crc kubenswrapper[4873]: E1201 09:54:29.432489 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:54:42 crc kubenswrapper[4873]: I1201 09:54:42.430916 4873 scope.go:117] "RemoveContainer" containerID="67abbbfa1eead10a18ea3d3104e50a98c59b5983a7e8d4be64c2d03f882521a7" Dec 01 09:54:42 crc kubenswrapper[4873]: E1201 09:54:42.431754 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:54:57 crc kubenswrapper[4873]: I1201 09:54:57.431165 4873 scope.go:117] "RemoveContainer" containerID="67abbbfa1eead10a18ea3d3104e50a98c59b5983a7e8d4be64c2d03f882521a7" Dec 01 09:54:57 crc kubenswrapper[4873]: E1201 09:54:57.432214 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:55:08 crc kubenswrapper[4873]: I1201 09:55:08.437746 4873 scope.go:117] "RemoveContainer" containerID="67abbbfa1eead10a18ea3d3104e50a98c59b5983a7e8d4be64c2d03f882521a7" Dec 01 09:55:08 crc kubenswrapper[4873]: E1201 09:55:08.438938 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:55:22 crc kubenswrapper[4873]: I1201 09:55:22.430883 4873 scope.go:117] "RemoveContainer" containerID="67abbbfa1eead10a18ea3d3104e50a98c59b5983a7e8d4be64c2d03f882521a7" Dec 01 09:55:22 crc kubenswrapper[4873]: E1201 09:55:22.432187 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:55:37 crc kubenswrapper[4873]: I1201 09:55:37.430448 4873 scope.go:117] "RemoveContainer" containerID="67abbbfa1eead10a18ea3d3104e50a98c59b5983a7e8d4be64c2d03f882521a7" Dec 01 09:55:37 crc kubenswrapper[4873]: E1201 09:55:37.431312 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:55:50 crc kubenswrapper[4873]: I1201 09:55:50.474973 4873 scope.go:117] "RemoveContainer" containerID="67abbbfa1eead10a18ea3d3104e50a98c59b5983a7e8d4be64c2d03f882521a7" Dec 01 09:55:50 crc kubenswrapper[4873]: E1201 09:55:50.478190 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:56:01 crc kubenswrapper[4873]: I1201 09:56:01.430608 4873 scope.go:117] "RemoveContainer" containerID="67abbbfa1eead10a18ea3d3104e50a98c59b5983a7e8d4be64c2d03f882521a7" Dec 01 09:56:01 crc kubenswrapper[4873]: E1201 09:56:01.431856 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:56:15 crc kubenswrapper[4873]: I1201 09:56:15.430651 4873 scope.go:117] "RemoveContainer" containerID="67abbbfa1eead10a18ea3d3104e50a98c59b5983a7e8d4be64c2d03f882521a7" Dec 01 09:56:15 crc kubenswrapper[4873]: E1201 09:56:15.431758 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:56:26 crc kubenswrapper[4873]: I1201 09:56:26.435705 4873 scope.go:117] "RemoveContainer" containerID="67abbbfa1eead10a18ea3d3104e50a98c59b5983a7e8d4be64c2d03f882521a7" Dec 01 09:56:26 crc kubenswrapper[4873]: E1201 09:56:26.437925 4873 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-scwpp_openshift-machine-config-operator(fef7b114-0e07-402d-a37b-315c36011f4b)\"" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" podUID="fef7b114-0e07-402d-a37b-315c36011f4b" Dec 01 09:56:37 crc kubenswrapper[4873]: I1201 09:56:37.430609 4873 scope.go:117] "RemoveContainer" containerID="67abbbfa1eead10a18ea3d3104e50a98c59b5983a7e8d4be64c2d03f882521a7" Dec 01 09:56:37 crc kubenswrapper[4873]: I1201 09:56:37.981331 4873 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-scwpp" event={"ID":"fef7b114-0e07-402d-a37b-315c36011f4b","Type":"ContainerStarted","Data":"f40a12f3ed3e331730c756213f0f330a3b59e6d90be3a4453ceafb0215ea0296"} var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515113263305024444 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015113263306017362 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015113251645016510 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015113251646015461 5ustar corecore